ramalama-chat - Man Page
OpenAI chat with the specified REST API URL
Synopsis
ramalama chat [options] [arg...]
positional arguments:
ARGS overrides the default prompt, and the output is
returned without entering the chatbot
Description
Chat with an OpenAI Rest API
Options
--api-key
OpenAI-compatible API key. Can also be set via the API_KEY environment variable.
--color
Indicate whether or not to use color in the chat. Possible values are "never", "always" and "auto". (default: auto)
--help, -h
Show this help message and exit
--list
List the available models at an endpoint
--model=MODEL
Model for inferencing (may not be required for endpoints that only serve one model)
--prefix
Prefix for the user prompt (default: 🦭 > )
--rag=path
A file or directory of files to be loaded and provided as local context in the chat history.
--url=URL
The host to send requests to (default: http://127.0.0.1:8080)
Examples
Communicate with the default local OpenAI REST API. (http://127.0.0.1:8080) With Podman containers.
$ ramalama chat 🦭 > Communicate with an alternative OpenAI REST API URL. With Docker containers. $ ramalama chat --url http://localhost:1234 🐋 >
See Also
History
Jun 2025, Originally compiled by Dan Walsh dwalsh@redhat.com ⟨mailto:dwalsh@redhat.com⟩