Jai Suphavadeeprasit
45f569f3af
clean
2026-03-18 09:20:08 -04:00
Jai Suphavadeeprasit
41947e98d6
clean
2026-03-17 12:25:38 -04:00
Jai Suphavadeeprasit
79baac1ea7
clean
2026-03-17 12:23:35 -04:00
Jai Suphavadeeprasit
805a0c0eac
revert to similar structure
2026-03-13 20:52:48 -04:00
Jai Suphavadeeprasit
1b8ff075c4
adding tests
2026-03-13 17:23:59 -04:00
Jai Suphavadeeprasit
697c594c72
changes
2026-03-13 16:58:37 -04:00
Jai Suphavadeeprasit
a8cdb53a4d
address problems
2026-03-13 16:12:05 -04:00
Jai Suphavadeeprasit
862cd3667d
clean logging
2026-03-13 12:38:52 -04:00
Jai Suphavadeeprasit
600c54f5f8
clean log
2026-03-13 12:12:33 -04:00
Jai Suphavadeeprasit
64794e7c72
sneaky bug
2026-03-13 11:06:00 -04:00
Jai Suphavadeeprasit
bb2736db4e
next
2026-03-13 11:05:40 -04:00
Jai Suphavadeeprasit
b91922082e
managed_Server pass through and centralize sem logic
2026-03-05 15:46:33 -05:00
pre-commit-ci[bot]
efc90bfb1b
[pre-commit.ci] auto fixes from pre-commit.com hooks
...
for more information, see https://pre-commit.ci
2026-03-04 04:18:12 +00:00
Jai Suphavadeeprasit
1eeb31065f
fixing comments
2026-03-03 23:16:05 -05:00
Jai Suphavadeeprasit
439b9b129b
prompt logprobs
2026-03-03 21:58:05 -05:00
Jai Suphavadeeprasit
b9291aa29f
init commit
2026-03-03 11:32:09 -05:00
Dakota
7d6aeb9bbf
add tokenizer name config to set the vllm/sglang tokenizer to something different if needed
2026-02-09 15:26:29 -06:00
Siddharth Balyan
7f28c52994
Merge branch 'main' into sid/verifiers
2026-01-16 11:50:27 +05:30
balyan.sid@gmail.com
6a27e88023
use managed server
2026-01-14 17:09:01 +05:30
teknium
e1ece3e64e
Add reasoning configuration support across server implementations
...
- Updated server classes (OpenAIServer, SGLangServer, TrlVllmServer, VLLMServer) to accept a ReasoningConfig parameter during initialization.
- Enhanced ReasoningConfig to allow flexible max_tokens without strict validation, accommodating varying provider limits.
- Implemented reasoning configuration injection in APIServer methods for chat and completion handling.
- Updated tests to reflect changes in max_tokens validation logic.
This commit integrates reasoning capabilities into the server handling architecture, improving compatibility with diverse reasoning models.
2026-01-05 23:20:01 +00:00
Dakota
e6ac3abdcb
add managed vllm server
2025-11-07 13:06:49 -06:00