ray.serve.llm.LLMRouter.chat#
- async LLMRouter.chat(body: ChatCompletionRequest) starlette.responses.Response #
Given a prompt, the model will return one or more predicted completions, and can also return the probabilities of alternative tokens at each position.
- Parameters:
body – The ChatCompletionRequest object.
- Returns:
A response object with completions.