tlm is an open-source command-line AI assistant designed to provide intelligent terminal support using locally running large language models. The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows developers to use powerful open-source models such as Llama, Phi, DeepSeek, and Qwen while maintaining privacy and avoiding external service dependencies. The system supports contextual queries where the AI analyzes files within a directory and generates answers based on project documentation or source code. It also detects the user’s shell environment automatically, allowing it to generate commands tailored to shells such as Bash, Zsh, or PowerShell.
Features
- Local AI terminal copilot powered by open-source language models
- Command generation based on natural language prompts
- Shell command explanation for Bash Zsh and PowerShell environments
- Context-aware question answering using project files and documentation
- Integration with Ollama for running models locally
- Interactive terminal mode for conversational assistance