Source code for ray.serve.llm.openai_api_models
from ray.llm._internal.serve.configs.openai_api_models import (
ChatCompletionRequest as _ChatCompletionRequest,
CompletionRequest as _CompletionRequest,
ChatCompletionStreamResponse as _ChatCompletionStreamResponse,
ChatCompletionResponse as _ChatCompletionResponse,
CompletionStreamResponse as _CompletionStreamResponse,
CompletionResponse as _CompletionResponse,
ErrorResponse as _ErrorResponse,
)
from ray.util.annotations import PublicAPI
[docs]
@PublicAPI(stability="alpha")
class ChatCompletionRequest(_ChatCompletionRequest):
"""ChatCompletionRequest is the request body for the chat completion API.
This model is compatible with vLLM's OpenAI API models.
"""
pass
[docs]
@PublicAPI(stability="alpha")
class CompletionRequest(_CompletionRequest):
"""CompletionRequest is the request body for the completion API.
This model is compatible with vLLM's OpenAI API models.
"""
pass
[docs]
@PublicAPI(stability="alpha")
class ChatCompletionStreamResponse(_ChatCompletionStreamResponse):
"""ChatCompletionStreamResponse is the response body for the chat completion API.
This model is compatible with vLLM's OpenAI API models.
"""
pass
[docs]
@PublicAPI(stability="alpha")
class ChatCompletionResponse(_ChatCompletionResponse):
"""ChatCompletionResponse is the response body for the chat completion API.
This model is compatible with vLLM's OpenAI API models.
"""
pass
[docs]
@PublicAPI(stability="alpha")
class CompletionStreamResponse(_CompletionStreamResponse):
"""CompletionStreamResponse is the response body for the completion API.
This model is compatible with vLLM's OpenAI API models.
"""
pass
[docs]
@PublicAPI(stability="alpha")
class CompletionResponse(_CompletionResponse):
"""CompletionResponse is the response body for the completion API.
This model is compatible with vLLM's OpenAI API models.
"""
pass
[docs]
@PublicAPI(stability="alpha")
class ErrorResponse(_ErrorResponse):
"""The returned response in case of an error."""
pass