add managed vllm server

This commit is contained in:
Dakota 2025-11-07 13:06:49 -06:00
parent 578175a709
commit e6ac3abdcb
9 changed files with 597 additions and 15 deletions

View file

@ -108,8 +108,8 @@ class ServerBaseline(BaseModel):
rolling_buffer_length: int = Field(
default=1000, description="Length of the rolling buffer to store metrics."
)
server_type: Literal["openai", "trl", "sglang"] = Field(
default="openai", description="Type of server to use, openai or trl"
server_type: Literal["openai", "trl", "sglang", "vllm"] = Field(
default="openai", description="Type of server to use"
)