• Install and run open-source models (almost all Hugging Face models)
• Package models with projects so environments are reproducible
• Serve models through an OpenAI compatible API
• Built-in chat UI for interacting with running models
• CLI workflow designed for developers
• Model benchmarking and rankings to help choose models Example:
llmpm serve llama3
This exposes an API endpoint similar to:
http://localhost:8080/v1/chat/completions
Would love to hear back from the community
dataversity•1h ago
llmpm install llama3
llmpm run llama3