Skip to content
View TechxGenus's full-sized avatar
🎯
Focusing
🎯
Focusing
  • USTC

Highlights

  • Pro

Block or report TechxGenus

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Starred repositories

Showing results

g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains

Python 2,848 235 Updated Sep 19, 2024
TypeScript 5,378 244 Updated Sep 19, 2024
Python 2,950 178 Updated Sep 20, 2024

A native PyTorch Library for large model training

Python 2,016 156 Updated Sep 19, 2024

A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 and reasoning techniques.

3,295 175 Updated Sep 20, 2024

[ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning

Jupyter Notebook 339 26 Updated Jun 29, 2024

EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Models (LLMs).

Jupyter Notebook 121 15 Updated Sep 11, 2024

Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793

Python 285 10 Updated Sep 18, 2024

[ICML 2024] TrustLLM: Trustworthiness in Large Language Models

Python 433 39 Updated Sep 6, 2024

🌟 Yi-Coder is a series of open-source code language models that delivers state-of-the-art coding performance with fewer than 10 billion parameters.

HTML 311 21 Updated Sep 18, 2024
Jupyter Notebook 126 5 Updated Sep 14, 2024

CUDA Templates for Linear Algebra Subroutines

C++ 5,380 907 Updated Sep 19, 2024

A throughput-oriented high-performance serving framework for LLMs

Cuda 482 17 Updated Sep 20, 2024

USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference

Python 314 20 Updated Sep 19, 2024

Official Implementation of EAGLE-1 and EAGLE-2

Python 754 74 Updated Aug 28, 2024

LinkedIn_AIHawk is a tool that automates the jobs application process on LinkedIn. Utilizing artificial intelligence, it enables users to apply for multiple job offers in an automated and personali…

Python 12,116 1,897 Updated Sep 17, 2024

LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.

Python 83 6 Updated Aug 23, 2024

Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads

Jupyter Notebook 2,213 150 Updated Jun 25, 2024

Efficient Triton Kernels for LLM Training

Python 2,966 152 Updated Sep 19, 2024

Code and data for "MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning" (ICLR 2024)

Jupyter Notebook 318 44 Updated Aug 25, 2024

Dafny is a verification-aware programming language

C# 2,880 256 Updated Sep 20, 2024

Lean 4 programming language and theorem prover

Lean 4,520 397 Updated Sep 20, 2024

🔨AI 方向好用的科研工具

2,302 344 Updated Jun 10, 2024

The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑‍🔬

Jupyter Notebook 7,534 1,005 Updated Sep 10, 2024

Microsoft Automatic Mixed Precision Library

Python 509 42 Updated Sep 18, 2024

Low-bit LLM inference on CPU with lookup table

C++ 443 32 Updated Sep 14, 2024

Helpful tools and examples for working with flex-attention

Python 349 14 Updated Aug 17, 2024

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilizatio…

Python 1,820 303 Updated Sep 19, 2024
Next