LLM CLI icon

LLM CLI

Command Line Assistants

Open-source CLI and Python library to run prompts, chat, embeddings, schemas, and tool-use across OpenAI, Claude, Gemini, and local models (Ollama, llama.cpp). Stores logs and vectors in SQLite and is extensible via plugins.

At a Glance

Pricing

Free tier available

Get started with LLM CLI at no cost with Apache 2.0 license and Full CLI and Python API.

Engagement

Available On

Windows
macOS
Linux
SDK

About LLM CLI

LLM is a lightweight command-line tool and Python library for working with large language models. It runs prompts and chat sessions, streams output, manages system prompts and templates, and logs everything to SQLite for later search and analysis. It supports structured outputs via JSON schemas, multimodal inputs (images/audio/video) through attachments, and can grant models controlled access to tools. Embeddings are first-class: you can generate, store, and run similarity search against vectors in SQLite. The plugin system adds providers and local runtimes—OpenAI, Anthropic (Claude), Google Gemini, Mistral, Ollama, llama.cpp, GPT4All and more—so you can mix remote APIs with models that run on your own machine. Installable with pip/pipx/Homebrew/uv, and usable as a Python API as well as a CLI. (Docs & feature list, plugins, CLI help, and PyPI requirements cited in sources.)

Demo Video

LLM CLI Demo Video
Watch on YouTube

Community Discussions

Be the first to start a conversation about LLM CLI

Share your experience with LLM CLI, ask questions, or help others learn from your insights.

Pricing

FREE

Free Plan Available

Get started with LLM CLI at no cost with Apache 2.0 license and Full CLI and Python API.

  • Apache 2.0 license
  • Full CLI and Python API
  • Plugin ecosystem
  • Community support
View official pricing

Capabilities

Key Features

  • Prompt execution with streaming output
  • Interactive chat mode (llm chat)
  • System prompts, templates, and fragments for long-context work
  • SQLite logging of prompts, responses, token usage, and metadata
  • Schema-based structured output (JSON) from models
  • Embeddings: generate, store, and similarity-search in SQLite
  • Multimodal attachments (image, audio, video) where models support it
  • Pluggable providers and local runtimes via llm install
  • Tool use / function calling support with safety notes
  • API key management (llm keys) and model discovery (llm models)
  • Python API parity for prompts, schemas, tools, fragments, and streaming
  • Configurable user content directory and custom locations

Integrations

OpenAI
Anthropic Claude
Google Gemini
Mistral AI
Ollama
llama.cpp / GGUF
GPT4All
Cohere
Groq
Replicate
OpenRouter
Azure OpenAI