Stars
The simplest way to serve AI/ML models in production
Evaluate your speech-to-text system with similarity measures such as word error rate (WER)
Reference-aware automatic speech evaluation toolkit
SOTA discrete acoustic codec models with 40 tokens per second for audio language modeling
Text-to-Music Generation with Rectified Flow Transformers
open-source multimodal large language model that can hear, talk while thinking. Featuring real-time end-to-end speech input and streaming audio output conversational capabilities.
💬 The most complete chat UI for React Native
SuperPrompt is an attempt to engineer prompts that might help us understand AI agents.
The open source Firebase alternative. Supabase gives you a dedicated Postgres database to build your web, mobile, and AI applications.
Create Reddit Videos with just✨ one command ✨
Accelerating the development of large multimodal models (LMMs) with lmms-eval
A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.
[CVPR 2024] Intelligent Grimm - Open-ended Visual Storytelling via Latent Diffusion Models
A lite request lib based on fetch with plugin support and similar API to axios.
This project collects GPU benchmarks from various cloud providers and compares them to fixed per token costs. Use our tool for efficient LLM GPU selections and cost-effective AI models. LLM provide…
Clean and simple starter repo using the T3 Stack along with Expo React Native and Supabase
Run PyTorch LLMs locally on servers, desktop and mobile
The official Python library for the Google Gemini API
VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)
Synthetic Minority Over-Sampling Technique for Regression
👔IMAGDressing👔: Interactive Modular Apparel Generation for Virtual Dressing
Actively maintained, community-driven chat UI implementation with an optional Firebase BaaS.
Instant voice cloning by MIT and MyShell.
MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory