Lists (1)
Sort Name ascending (A-Z)
Stars
A reading list on LLM based Synthetic Data Generation 🔥
Python fast on-disk dictionary / RocksDB & SpeeDB Python binding
open-source multimodal large language model that can hear, talk while thinking. Featuring real-time end-to-end speech input and streaming audio output conversational capabilities.
Typer, build great CLIs. Easy to code. Based on Python type hints.
Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 40+ benchmarks
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
Official inference repo for FLUX.1 models
Classical equations and diagrams in machine learning
Utils for streaming large files (S3, HDFS, gzip, bz2...)
Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verified research papers.
🐚 Python-powered shell. Full-featured and cross-platform.
Official implementation of ⚡ Flash Diffusion ⚡: Accelerating Any Conditional Diffusion Model for Few Steps Image Generation
🔥LeetCode solutions in any programming language | 多种编程语言实现 LeetCode、《剑指 Offer(第 2 版)》、《程序员面试金典(第 6 版)》题解
A guidance language for controlling large language models.
TerminalTextEffects (TTE) is a terminal visual effects engine, application, and Python library.
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) avai…
CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts
[ICML 2024] CLLMs: Consistency Large Language Models
UpTrain is an open-source unified platform to evaluate and improve Generative AI applications. We provide grades for 20+ preconfigured checks (covering language, code, embedding use-cases), perform…
[ACL 2024 Demo] Official GitHub repo for UltraEval: An open source framework for evaluating foundation models.
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
MiniCPM3-4B: An edge-side LLM that surpasses GPT-3.5-Turbo.
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
Accelerating the development of large multimodal models (LMMs) with lmms-eval
Reaching LLaMA2 Performance with 0.1M Dollars