LM Studio
The most comprehensive local LLM desktop application. Discover, download, and chat with models through a polished UI with built-in OpenAI-compatible API server.
LM Studio is the most popular desktop application for running large language models locally. It provides a complete graphical environment for discovering models on Hugging Face, downloading quantized variants, configuring inference parameters, and chatting with models — all without touching a terminal. For users who want powerful local AI with a visual interface, LM Studio is the standard choice.
Key Features
Integrated model discovery. LM Studio connects directly to Hugging Face and presents a searchable catalog of compatible models. Each listing shows available quantization levels, file sizes, and RAM requirements, making it straightforward to find models that fit your hardware. One-click downloads handle the rest.
Visual chat interface. The chat UI supports markdown rendering, code highlighting, conversation branching, and side-by-side model comparison. You can adjust temperature, top-p, repeat penalty, context length, and other parameters in real time through sliders and input fields.
Local API server. LM Studio includes a built-in server that exposes an OpenAI-compatible API on a configurable port. This turns your desktop into a local AI endpoint that works with any tool expecting the OpenAI format — VS Code extensions, web applications, Python scripts, and more.
Hardware optimization. LM Studio automatically detects your GPU and configures layer offloading. It supports NVIDIA GPUs via CUDA, AMD GPUs via Vulkan, and Apple Silicon via Metal. The interface displays real-time metrics for tokens per second, memory usage, and GPU utilization.
Multi-model management. Keep multiple models downloaded and switch between them instantly. LM Studio tracks your conversation history per model and lets you organize chats into folders.
When to Use LM Studio
Choose LM Studio when you prefer a graphical interface over command-line tools, want to quickly compare different models and settings, or need a local API server without writing any configuration. It is particularly well-suited for non-technical users, researchers evaluating models, and developers who want a visual playground before committing to a specific model for their application.
Ecosystem Role
LM Studio uses llama.cpp under the hood and reads the same GGUF model files. It complements command-line tools like Ollama by offering a richer visual experience. For team or multi-user deployments, pair LM Studio’s API server with a frontend like Open WebUI.