Skip to content

LLMs & Chat Models

Large language models, chat assistants, and the runtimes that serve them. From flagship hosted models like Claude, GPT, and Gemini to local-first runners like Ollama, LM Studio, vLLM, and Llamafile. As the model landscape evolves we'll keep growing this category.

LM Studio Logo

LM Studio

Desktop app to run AI models locally and privately. Headless server mode, JS/Python SDKs, supports gpt-oss, Qwen, Gemma, DeepSeek.

AI AssistantSDKLLMLocal LLM
Jan Logo

Jan

Open-source ChatGPT-style desktop assistant with 4M+ downloads. Personal intelligence that answers only to you.

AI AssistantLLMLocal LLMOpen Source
vLLM Logo

vLLM

High-throughput, memory-efficient open-source LLM inference and serving engine. Production-grade, originally from UC Berkeley.

InfrastructureLLMOpen SourceAPI
LocalAI Logo

LocalAI

OpenAI/Anthropic-compatible REST API for running LLMs, agents, and document intelligence locally on your hardware.

LLMLocal LLMOpen SourceAPI
Text Generation WebUI Logo

Text Generation WebUI

The original Gradio web UI for running local LLMs. 100% private, 100% offline, multi-backend, multimodal.

LLMLocal LLMOpen SourceText Generation
Llamafile Logo

Llamafile

Mozilla.ai project that distributes and runs LLMs as single-file executables across any OS using Cosmopolitan Libc.

LLMLocal LLMOpen Source

6 tools in this category