Home

Awesome

Awesome Local AI Tweet

If you tried Jan Desktop and liked it, please also check out the following awesome collection of open source and/or local AI tools and solutions.

Your contributions are always welcome!

Lists

Inference Engine

RepositoryDescriptionSupported model formatsCPU/GPU SupportUIlanguagePlatform Type
llama.cpp- Inference of LLaMA model in pure C/C++GGML/GGUFBothC/C++Text-Gen
Cortex- Multi-engine engine embeddable in your apps. Uses llama.cpp and moreBothBothText-Gen
ollama- CLI and local server. Uses llama.cppBothBothText-Gen
koboldcpp- A simple one-file way to run various GGML models with KoboldAI's UIGGMLBothC/C++Text-Gen
LoLLMS- Lord of Large Language Models Web User Interface.Nearly ALLBothPythonText-Gen
ExLlama- A more memory-efficient rewrite of the HF transformers implementation of LlamaAutoGPTQ/GPTQGPUPython/C++Text-Gen
vLLM- vLLM is a fast and easy-to-use library for LLM inference and serving.GGML/GGUFBothPythonText-Gen
SGLang- 3-5x higher throughput than vLLM (Control flow, RadixAttention, KV cache reuse)Safetensor / AWQ / GPTQGPUPythonText-Gen
LmDeploy- LMDeploy is a toolkit for compressing, deploying, and serving LLMs.Pytorch / TurbomindBothPython/C++Text-Gen
Tensorrt-llm- Inference efficiently on NVIDIA GPUsPython / C++ runtimesBothPython/C++Text-Gen
CTransformers- Python bindings for the Transformer models implemented in C/C++ using GGML libraryGGML/GPTQBothC/C++Text-Gen
llama-cpp-python- Python bindings for llama.cppGGUFBothPythonText-Gen
llama2.rs- A fast llama2 decoder in pure RustGPTQCPURustText-Gen
ExLlamaV2- A fast inference library for running LLMs locally on modern consumer-class GPUsGPTQ/EXL2GPUPython/C++Text-Gen
LoRAX- Multi-LoRA inference server that scales to 1000s of fine-tuned LLMsSafetensor / AWQ / GPTQGPUPython/RustText-Gen
text-generation-inference- Inference serving toolbox with optimized kernels for each LLM architectureSafetensors / AWQ / GPTQBothPython/RustText-Gen

Inference UI

Platforms / full solutions

Developer tools

User Tools

Agents

Training

LLM Leaderboard

Research

Community