llmpm
LLM PACKAGE MANAGER
Download and run 10,000+ models from HuggingFace, Ollama & Mistral with a single command.
$
pip install llmpm$
npm install -g llmpm100K+
Models Available
3
Model Sources
5
Model Types
1
Command to Start
INSTALLATION
Install llmpm in seconds. Lightweight CLI — backends installed automatically on first run.
sh
#Recommended: installs CLI + ML backends into isolated venv$pip install llmpm$#Selective backends$pip install llmpm[gguf] # GGUF / llama.cpp$pip install llmpm[transformers] # text generation$pip install llmpm[diffusion] # image generation
✓ Isolated venv at ~/.llmpm/venv — no system Python pollution
QUICK START
From install to running a model in under a minute.
sh
#Install a model$llmpm install meta-llama/Llama-3.2-3B-Instruct$#Run it interactively$llmpm run meta-llama/Llama-3.2-3B-Instruct$#Or serve as OpenAI-compatible API$llmpm serve meta-llama/Llama-3.2-3B-Instruct$#Browse installed models$llmpm list
COMMANDS
Everything you need to manage large language models.
Install Models
Install any model from HuggingFace, Ollama, or Mistral with a single command.
$
llmpm install meta-llama/Llama-3.2-3B-InstructRun Instantly
Auto-detects model type and launches the right backend. Text, image, audio — it just works.
$
llmpm run meta-llama/Llama-3.2-3B-InstructServe as API
Expose any model as an OpenAI-compatible REST API with a built-in chat UI.
$
llmpm serve meta-llama/Llama-3.2-3B-InstructPush Models
Upload fine-tuned models directly to HuggingFace Hub with one command.
$
llmpm push my-org/my-fine-tuneSUPPORTED MODEL TYPES
One tool for every modality.
💬
Text Generation
GGUF via llama.cpp & Transformer checkpoints
🎨
Text-to-Image
Diffusion models via Diffusers
👁️
Vision Models
Image-to-text, captioning & VQA
🎤
Speech-to-Text
ASR with Whisper and more
🔊
Text-to-Speech
TTS generation to WAV output
Start in 30 seconds.
No config files, no complex setup. Just install and run.