Skip to content

chat: Conversational Generation

fastdeploy chat interacts with a running API server to generate chat responses.

Parameters

Parameter Description Default
--url URL of the running OpenAI-compatible RESTful API server http://localhost:8000/v1
--model-name Name of the model to use for prompt completion; defaults to the first model listed in the models API None
--api-key API key for OpenAI services; overrides environment variable if provided None
--system-prompt Specifies the system prompt used in the chat template None
-q, --quick Sends a single prompt as a MESSAGE, prints the response, and exits None

Examples

# Connect directly to a local API
fastdeploy chat

# Specify an API URL
fastdeploy chat --url http://{fastdeploy-serve-host}:{fastdeploy-serve-port}/v1

# Send a single quick prompt
fastdeploy chat --quick "hi"