chat: Conversational Generation
fastdeploy chat interacts with a running API server to generate chat responses.
Parameters
| Parameter | Description | Default |
|---|---|---|
| --url | URL of the running OpenAI-compatible RESTful API server | http://localhost:8000/v1 |
| --model-name | Name of the model to use for prompt completion; defaults to the first model listed in the models API | None |
| --api-key | API key for OpenAI services; overrides environment variable if provided | None |
| --system-prompt | Specifies the system prompt used in the chat template | None |
| -q, --quick | Sends a single prompt as a MESSAGE, prints the response, and exits | None |
Examples
# Connect directly to a local API
fastdeploy chat
# Specify an API URL
fastdeploy chat --url http://{fastdeploy-serve-host}:{fastdeploy-serve-port}/v1
# Send a single quick prompt
fastdeploy chat --quick "hi"