Source code for ray.serve.llm.openai_api_models

from ray.llm._internal.serve.configs.openai_api_models import (
    ChatCompletionRequest as _ChatCompletionRequest,
    CompletionRequest as _CompletionRequest,
    ChatCompletionStreamResponse as _ChatCompletionStreamResponse,
    ChatCompletionResponse as _ChatCompletionResponse,
    CompletionStreamResponse as _CompletionStreamResponse,
    CompletionResponse as _CompletionResponse,
    ErrorResponse as _ErrorResponse,
)

from ray.util.annotations import PublicAPI


[docs] @PublicAPI(stability="alpha") class ChatCompletionRequest(_ChatCompletionRequest): """ChatCompletionRequest is the request body for the chat completion API. This model is compatible with vLLM's OpenAI API models. """ pass
[docs] @PublicAPI(stability="alpha") class CompletionRequest(_CompletionRequest): """CompletionRequest is the request body for the completion API. This model is compatible with vLLM's OpenAI API models. """ pass
[docs] @PublicAPI(stability="alpha") class ChatCompletionStreamResponse(_ChatCompletionStreamResponse): """ChatCompletionStreamResponse is the response body for the chat completion API. This model is compatible with vLLM's OpenAI API models. """ pass
[docs] @PublicAPI(stability="alpha") class ChatCompletionResponse(_ChatCompletionResponse): """ChatCompletionResponse is the response body for the chat completion API. This model is compatible with vLLM's OpenAI API models. """ pass
[docs] @PublicAPI(stability="alpha") class CompletionStreamResponse(_CompletionStreamResponse): """CompletionStreamResponse is the response body for the completion API. This model is compatible with vLLM's OpenAI API models. """ pass
[docs] @PublicAPI(stability="alpha") class CompletionResponse(_CompletionResponse): """CompletionResponse is the response body for the completion API. This model is compatible with vLLM's OpenAI API models. """ pass
[docs] @PublicAPI(stability="alpha") class ErrorResponse(_ErrorResponse): """The returned response in case of an error.""" pass