- All languages
- ASP
- Assembly
- Batchfile
- BitBake
- C
- C#
- C++
- CMake
- CSS
- CoffeeScript
- Cuda
- Cython
- D
- Dart
- Dockerfile
- Fortran
- GLSL
- Go
- HLSL
- HTML
- Handlebars
- Haskell
- Java
- JavaScript
- JetBrains MPS
- Jsonnet
- Julia
- Jupyter Notebook
- Kotlin
- LLVM
- Lua
- MATLAB
- MDX
- MLIR
- Makefile
- Markdown
- PHP
- Pascal
- Perl
- PowerShell
- PureBasic
- Python
- Riot
- Ruby
- Rust
- Sass
- Scala
- Shell
- Smarty
- TeX
- TypeScript
- Verilog
- Vim Script
- Vue
Starred repositories
🎧 Open source Spotify client that doesn't require Premium nor uses Electron! Available for both desktop & mobile!
🍃 Organic Maps is a free Android & iOS offline maps app for travelers, tourists, hikers, and cyclists. It uses crowd-sourced OpenStreetMap data and is developed with love by MapsWithMe (MapsMe) fou…
A lightweight library for portable low-level GPU computation using WebGPU.
A modular graph-based Retrieval-Augmented Generation (RAG) system
C++ template library for high performance SIMD based sorting algorithms
Backward compatible ML compute opset inspired by HLO/MHLO
LightSeq: A High Performance Library for Sequence Processing and Generation
A Python framework for high performance GPU simulation and graphics
Develop Desktop, Embedded, Mobile and WebAssembly apps with C# and XAML. The most popular .NET UI client technology
一个还算强大的Web思维导图。A relatively powerful web mind map.
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantization https://arxiv.org/pdf/2401.06118.pdf and PV-Tuning: Beyond Straight-Through Estimation for Ext…
A fast inference library for running LLMs locally on modern consumer-class GPUs
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
Advanced Quantization Algorithm for LLMs. This is official implementation of "Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs"
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
A comprehensive library for implementing LLMs, including a unified training pipeline and comprehensive model evaluation.
A framework for few-shot evaluation of language models.
Fast and memory-efficient exact attention
Hackable and optimized Transformers building blocks, supporting a composable construction.
Ongoing research training transformer models at scale