ray.serve.llm.LLMRouter.chat#
- async LLMRouter.chat(body: ChatCompletionRequest) starlette.responses.Response#
 Given a prompt, the model will return one or more predicted completions, and can also return the probabilities of alternative tokens at each position.
- Parameters:
 body – The ChatCompletionRequest object.
- Returns:
 A response object with completions.