llmpm

LLM PACKAGE MANAGER

Download and run 10,000+ models from HuggingFace, Ollama & Mistral with a single command.

$pip install llmpm
$npm install -g llmpm
100K+
Models Available
3
Model Sources
5
Model Types
1
Command to Start

INSTALLATION

Install llmpm in seconds. Lightweight CLI — backends installed automatically on first run.

sh
#Recommended: installs CLI + ML backends into isolated venv
$pip install llmpm
$
#Selective backends
$pip install llmpm[gguf] # GGUF / llama.cpp
$pip install llmpm[transformers] # text generation
$pip install llmpm[diffusion] # image generation

✓ Isolated venv at ~/.llmpm/venv — no system Python pollution

QUICK START

From install to running a model in under a minute.

sh
#Install a model
$llmpm install meta-llama/Llama-3.2-3B-Instruct
$
#Run it interactively
$llmpm run meta-llama/Llama-3.2-3B-Instruct
$
#Or serve as OpenAI-compatible API
$llmpm serve meta-llama/Llama-3.2-3B-Instruct
$
#Browse installed models
$llmpm list

COMMANDS

Everything you need to manage large language models.

Install Models

Install any model from HuggingFace, Ollama, or Mistral with a single command.

$llmpm install meta-llama/Llama-3.2-3B-Instruct

Run Instantly

Auto-detects model type and launches the right backend. Text, image, audio — it just works.

$llmpm run meta-llama/Llama-3.2-3B-Instruct

Serve as API

Expose any model as an OpenAI-compatible REST API with a built-in chat UI.

$llmpm serve meta-llama/Llama-3.2-3B-Instruct

Push Models

Upload fine-tuned models directly to HuggingFace Hub with one command.

$llmpm push my-org/my-fine-tune

SUPPORTED MODEL TYPES

One tool for every modality.

💬
Text Generation
GGUF via llama.cpp & Transformer checkpoints
🎨
Text-to-Image
Diffusion models via Diffusers
👁️
Vision Models
Image-to-text, captioning & VQA
🎤
Speech-to-Text
ASR with Whisper and more
🔊
Text-to-Speech
TTS generation to WAV output

Start in 30 seconds.

No config files, no complex setup. Just install and run.