vllm chat¶
Arguments¶
--url¶
url of the running OpenAI-Compatible RESTful API server
Default: http://localhost:8000/v1
--model-name¶
The model name used in prompt completion, default to the first model in list models API call.
Default: None
--api-key¶
API key for OpenAI services. If provided, this api key will overwrite the api key obtained through environment variables. It is important to note that this option only applies to the OpenAI-compatible API endpoints and NOT other endpoints that may be present in the server. See the security guide in the vLLM docs for more details.
Default: None
--system-prompt¶
The system prompt to be added to the chat template, used for models that support system prompts.
Default: None
-q, --quick¶
Send a single prompt as MESSAGE and print the response, then exit.
Default: None