oterm
the text-based terminal client for Ollama.
Features
- intuitive and simple terminal UI, no need to run servers, frontends, just type
oterm
in your terminal. - multiple persistent chat sessions, stored together with system prompt & parameter customizations in sqlite.
- support for Model Context Protocol (MCP) tools & prompts integration.
- can use any of the models you have pulled in Ollama, or your own custom models.
- allows for easy customization of the model's system prompt and parameters.
- supports tools integration for providing external information to the model.
Installation
See the Installation section.
Using oterm
In order to use oterm
you will need to have the Ollama server running. By default it expects to find the Ollama API running on http://127.0.0.1:11434
. If you are running Ollama inside docker or on a different host/port, use the OLLAMA_HOST
environment variable to customize the host/port. Alternatively you can use OLLAMA_URL
to specify the full http(s) url. Setting OTERM_VERIFY_SSL
to False
will disable SSL verification.
To start oterm
simply run:
Screenshots
The splash screen animation that greets users when they start oterm.
A view of the chat interface, showcasing the conversation between the user and the model.
oterm supports multiple themes, allowing users to customize the appearance of the interface.
License
This project is licensed under the MIT License.