Skip to main content
LM Studio provides a user-friendly way to run large language models locally on your computer, offering privacy, speed, and offline capabilities without requiring an internet connection.

Overview

LM Studio bridges the gap between powerful AI models and local computing, allowing you to run advanced AI models directly on your machine. It’s perfect for users who want privacy, speed, and control over their AI interactions.

Local Execution

Run AI models directly on your computer

Privacy First

Keep conversations and data completely private

Offline Capable

Work without internet connectivity

How It Works

LM Studio downloads and runs AI models locally using your computer’s resources. It provides a simple interface to manage models, start local servers, and connect to various applications including Codinit.

Downloading Models

Choose from thousands of available models in various sizes and capabilities.
  • Model Library: Browse and download models from Hugging Face
  • Size Options: From small 1GB models to large 100GB+ models
  • Format Support: GGUF, SafeTensor, and other formats
  • Automatic Updates: Stay current with latest model versions

Running AI Locally

Start a local API server that applications can connect to.
  • One-Click Setup: Start local server with single button
  • API Compatibility: OpenAI-compatible API endpoints
  • Multi-Platform: Windows, macOS, and Linux support
  • Resource Management: Monitor CPU/GPU usage and memory

Optimization Settings

Fine-tune performance based on your hardware capabilities.
  • GPU Acceleration: Utilize NVIDIA/AMD GPUs when available
  • CPU Optimization: Efficient CPU inference for all systems
  • Memory Management: Control RAM usage and model loading
  • Quantization: Balance speed vs. quality with different precisions

Setup Instructions

1

Download LM Studio

Visit LM Studio website and download the application
2

Install and Launch

Install LM Studio and launch the application
3

Download Models

Browse the model library and download models you want to use
4

Start Local Server

Click “Start Server” in LM Studio to begin the local API server
5

Configure in Codinit

Set the server URL (usually http://localhost:1234) in Codinit settings
6

Test Connection

Verify the connection and start using local AI models

Key Features

Platform Advantages

  • Complete Privacy: All conversations stay on your device
  • No API Costs: Run unlimited AI interactions for free
  • Offline Operation: Work without internet connectivity
  • Hardware Flexibility: Run on any modern computer
  • Model Variety: Access thousands of different AI models

Use Cases

Secure Development

Perfect for sensitive development work and private projects.
  • Code review without sharing code externally
  • Private documentation and analysis
  • Secure brainstorming and planning
  • Confidential business applications

Offline Productivity

Continue working with AI assistance even without internet.
  • Travel and remote work scenarios
  • Limited connectivity environments
  • Data-sensitive offline processing
  • Emergency backup AI capabilities

Budget-Friendly AI

Access advanced AI capabilities without ongoing costs.
  • Unlimited usage without API fees
  • No per-token or per-request charges
  • One-time setup, ongoing free usage
  • Cost-effective for heavy AI users

Educational Use

Learn about AI and experiment with different models.
  • Study different model architectures
  • Compare model performance and capabilities
  • Learn prompt engineering techniques
  • Understand AI model behaviors

System Requirements

Basic Setup

Requirements for running small to medium models.
  • RAM: 8GB minimum, 16GB recommended
  • Storage: 10GB free space for models and application
  • OS: Windows 10+, macOS 10.15+, Ubuntu 18.04+
  • CPU: Modern multi-core processor

Hardware Acceleration

Utilize GPU acceleration for faster inference speeds.
  • NVIDIA GPUs: CUDA support for maximum performance
  • AMD GPUs: ROCm support on Linux
  • Apple Silicon: Native acceleration on M1/M2/M3 Macs
  • CPU Fallback: Automatic fallback to CPU when GPU unavailable

Model Selection Guide

Choosing Model Size

Balance between performance and resource requirements.
  • Small Models (1-3GB): Fast, basic capabilities, good for simple tasks
  • Medium Models (3-7GB): Balanced performance, good for most applications
  • Large Models (7-20GB): High quality, slower but more capable
  • XL Models (20GB+): Maximum quality, requires powerful hardware

Specialized Models

Choose models based on your specific needs.
  • Code Models: Code generation, debugging, technical writing
  • General Chat: Conversation, analysis, creative writing
  • Math/Science: Mathematical reasoning, scientific analysis
  • Multilingual: Support for multiple languages and cultures
Free Forever: LM Studio is completely free to use. No subscriptions or hidden costs.
Start Small: Begin with smaller models to test your setup, then upgrade to larger models as needed.
Resource Intensive: Large models require significant RAM and may run slowly on lower-end hardware.