Commit graph

29 commits

Author SHA1 Message Date
ansulx
d97f366ae0 Add regression test for TRL vLLM completion wrapper
Ensure the TRL vLLM completion wrapper returns a Completion with text so issue #183 stays covered.
2026-02-06 01:57:16 +05:30
VolodymyrBg
77a3505955
Update test_openai_api_workarounds.py 2026-01-29 10:13:50 +02:00
Teknium
84a8bbb9cb
Merge pull request #317 from Savage890/fix/issue-308-jsonl2html
fix: handle nested message format in jsonl2html.py
2026-01-16 06:47:44 -08:00
teknium
31a8cdc7a7 update test to reflect the change in reasoning effort mapping 2026-01-15 07:48:52 +00:00
teknium
0316cac8d1 Rename is_active method to is_reasoning_kwargs_active in ReasoningConfig for clarity. Update references in the class and corresponding tests to reflect this change. 2026-01-15 06:26:31 +00:00
pre-commit-ci[bot]
39e9a233db [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
2026-01-14 21:36:05 +00:00
Savage890
39f05d18fa fix: handle nested message format in jsonl2html.py (#308) 2026-01-15 03:01:15 +05:30
pre-commit-ci[bot]
6cfcbdf4d5 [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
2026-01-05 23:20:47 +00:00
teknium
e1ece3e64e Add reasoning configuration support across server implementations
- Updated server classes (OpenAIServer, SGLangServer, TrlVllmServer, VLLMServer) to accept a ReasoningConfig parameter during initialization.
- Enhanced ReasoningConfig to allow flexible max_tokens without strict validation, accommodating varying provider limits.
- Implemented reasoning configuration injection in APIServer methods for chat and completion handling.
- Updated tests to reflect changes in max_tokens validation logic.

This commit integrates reasoning capabilities into the server handling architecture, improving compatibility with diverse reasoning models.
2026-01-05 23:20:01 +00:00
teknium
89c9697665 fix test 2025-12-30 23:08:54 +00:00
teknium
127a925471 Merge branch 'add_reasoning_handling_draft' of https://github.com/NousResearch/atropos into add_reasoning_handling_draft 2025-12-30 11:59:46 +00:00
teknium
747fbc9285 fix linting 2025-12-30 11:56:21 +00:00
pre-commit-ci[bot]
97047eee7b [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
2025-12-30 00:26:33 +00:00
teknium
62fa51240c Add support for reasoning models and their variety of providers/endpoints 2025-12-30 00:23:00 +00:00
dmahan93
b1e164eef5
Merge pull request #264 from NousResearch/add-logprob-server-manager-fn
add sglang specific token level logprob handling and server manager/b…
2025-10-29 13:53:39 -07:00
Dakota
3c8fc32288 fix test case 2025-10-29 14:38:16 -05:00
pre-commit-ci[bot]
0d80da5146 [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
2025-10-24 20:10:29 +00:00
dmahan93
7bf4cfbf80 add managed server to make grabbing logprobs easier w/ tokenized items 2025-10-24 13:09:46 -07:00
pre-commit-ci[bot]
0840c26e94 [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
2025-10-15 04:19:25 +00:00
ropresearch
e5b8fb8654 clean up 2025-10-10 11:50:39 -04:00
ropresearch
baf4b2d8a8 gzip compression for atropos api 2025-10-10 01:26:52 -04:00
ropresearch
c3fc68879c group temps, sample temps, and logprob api params 2025-09-25 16:41:58 -04:00
Dakota
08e14cc745 feat: add minimum batch allocation support for environments
- Add min_batch_allocation parameter to ensure environments contribute minimum proportion to each batch
- Implement grab_batch_with_minimum_allocations function with proper scaling when allocations exceed 100%
- Add mixed-size group buffering to handle variable-sized data submissions
- Update server to use minimum allocation logic when any env has min_batch_allocation set
- Add comprehensive tests for minimum allocation scenarios
- Update documentation in API README and CONFIG.md
- Update example environments to demonstrate the feature

This feature allows critical environments to guarantee they contribute at least a specified proportion (0.0-1.0) to each training batch, ensuring important data sources are always represented during training.

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-07-07 08:50:28 -05:00
Dakota
e13526d308 Fix API to accept messages without reward field + comprehensive tests
- Made reward field truly optional in messages (no auto-addition)
- Accept custom roles (dog, cat, etc.) beyond standard ones
- Added 24 new tests for edge cases (tuples, unicode, large content)
- Reorganized test structure: moved from testing/ to atroposlib/tests/
- Fixed legacy API tests and removed tests requiring missing data files

All 43 tests pass\! Fixes message handling for SFT use cases.

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-06-09 14:03:08 -05:00
dmahan93
96be544228 Merge commit '71e7a5ca27' into add-support-for-custom-api-servers 2025-05-12 18:40:35 -05:00
dmahan93
71e7a5ca27
Merge pull request #41 from NousResearch/workaround-provider-ignoring-n-kwarg-openai-api
Add n kwarg being ignored workaround
2025-05-12 18:19:47 -05:00
dmahan93
1aa72d7e7e Add n kwarg being ignored workaround 2025-05-12 12:06:03 -05:00
dmahan93
727c7ba640 Remove dependency on torch for default installation 2025-05-12 10:17:41 -05:00
Dakota Nous
621d00dd80 first commit 2025-04-29 12:10:10 -07:00