Inference Runtimes

🗺 Open the interactive Open-Source AI map →
Visualize the full ecosystem of open-source models, runtimes, frameworks, and agents.

Inference engines and runtimes — Ollama, vLLM, llama.cpp, Text Generation Inference (TGI), LM Studio, LocalAI. Performance benchmarks, GPU/CPU optimization, quantization.