Commit graph

10 commits

Author SHA1 Message Date
teknium
e1ece3e64e Add reasoning configuration support across server implementations
- Updated server classes (OpenAIServer, SGLangServer, TrlVllmServer, VLLMServer) to accept a ReasoningConfig parameter during initialization.
- Enhanced ReasoningConfig to allow flexible max_tokens without strict validation, accommodating varying provider limits.
- Implemented reasoning configuration injection in APIServer methods for chat and completion handling.
- Updated tests to reflect changes in max_tokens validation logic.

This commit integrates reasoning capabilities into the server handling architecture, improving compatibility with diverse reasoning models.
2026-01-05 23:20:01 +00:00
pre-commit-ci[bot]
312f8859e3 [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
2025-10-16 18:47:07 +00:00
Dakota
134cbc09d0 update openai/trl_vllm server with new fn 2025-10-16 13:45:55 -05:00
Shannon Sands
76736ec4c2 removed unused 2025-05-17 13:37:55 -07:00
Shannon Sands
c05d9f7f53 added health check flag to skip entirely 2025-05-17 13:36:26 -07:00
Shannon Sands
90138376f9 thanks cursor 2025-05-17 13:00:35 -07:00
Shannon Sands
59b08a1aa9 Added flag for chat completion, switch in OpenAI client 2025-05-17 12:59:20 -07:00
dmahan93
df62979b90 refactor to not mess up process... 2025-05-13 09:22:07 -05:00
dmahan93
6e9405ba95 Fix bad merge 2025-05-12 20:02:54 -05:00
dmahan93
0aaf59fc9a add trl server
add gsm8k example for axolotl checking
2025-05-12 19:04:46 -05:00