Highlights
Lists (8)
Sort Name ascending (A-Z)
Stars
【大模型】3小时完全从0训练一个仅有26M的小参数GPT,最低仅需2G显卡即可推理训练!
A collection of awesome video generation studies.
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
A high-throughput and memory-efficient inference and serving engine for LLMs
Efficient Triton Kernels for LLM Training
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Evaluation framework for your Retrieval Augmented Generation (RAG) pipelines
SGLang is a fast serving framework for large language models and vision language models.
Scalable toolkit for efficient model alignment
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
A modular graph-based Retrieval-Augmented Generation (RAG) system
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
Example models using DeepSpeed
Get up and running with Llama 3.1, Mistral, Gemma 2, and other large language models.
User-friendly WebUI for LLMs (Formerly Ollama WebUI)
A generative speech model for daily dialogue.
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
[ICML 2024] Official repository of the paper: "Diving into Underwater: Segment Anything Model Guided Underwater Salient Instance Segmentation and A Large-scale Dataset"
A library for building fast, reliable and evolvable network services.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
《神经网络与深度学习》 邱锡鹏著 Neural Network and Deep Learning
《Designing Data-Intensive Application》DDIA中文翻译