Back to Library
Local LLM

LM Studio

LM Studio application screenshot

Your Local AI Toolkit

LM Studio allows you to easily download and run open-source large language models like Llama, DeepSeek, Qwen, and Phi locally on your computer. It provides a beginner-friendly interface with expert features.

Key Features

  • Run LLMs Locally: Use powerful models right on your laptop or PC
  • Discover Models: Find and download open-source models through the built-in catalog
  • Local LLM Server: Power your applications with a local API server
  • Chat with Documents: Implement RAG (Retrieval-Augmented Generation) with your local files
  • Cross-platform SDK: Build local AI apps with Python or TypeScript libraries

Getting Started

  1. Download LM Studio for your platform (Windows, macOS, Linux)
  2. Browse and download open-source models
  3. Start chatting or run a local LLM server
  4. Integrate with applications using the SDK

Perfect For

  • Developers building AI-powered applications
  • Privacy-focused users who want to keep data local
  • Anyone exploring AI without cloud costs
  • Researchers experimenting with different models

LM Studio is designed for privacy with no data collection, keeping everything local to your machine.

Apple silicon as-a-Service

Discover why Macfleet is the preferred cloud provider for developers.