update to tech report version (#10)

* feat(run_eval): add checkpoint resume functionality and update example documentation;
- update new bootcamp benchmark dataset

* refactor(data_pipeline): optimize data generation pipeline; add multiple preset configurations for data generation

* docs: update bootcamp list and add new scripts

- Update Fulllist_InternBootcamp.md with new bootcamps and categories
- Add new scripts to .gitignore:
  - examples/pipelines/filter_autogen_configs.py
  - examples/pipelines/quickgen_data_configs_from_eval_meta.py
- Update dependencies in setup.py:
  - Add scipy and scikit-learn

* refactor(internbootcamp): update bootcamp modules and improve error handling

- Update import statements in __init__.py files
- Add timestamp to target directory name in verl_data_preprocess.py
- Improve error handling and scoring logic in bootcamp_judger.py
- Remove unnecessary comments and update puzzle descriptions in multiple files
This commit is contained in:
Yongkang Chen 2025-08-28 12:39:47 +08:00 committed by GitHub
parent 125a7818e0
commit a8249acc18
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2952 changed files with 105460 additions and 17649 deletions

View file

@ -47,7 +47,7 @@ class Basebootcamp:
@classmethod
def verify_score(cls, model_output, identity: dict, format_score=0, short_penalty=False, short_threshold=256, think_threshold=128, ans_threshold=128, format_penalty=False) -> float:
def verify_score(cls, model_output, identity: dict, format_score=0, short_penalty=False, short_threshold=512, think_threshold=256, ans_threshold=256, format_penalty=False) -> float:
"""
Verify the output against the ground truth.
@ -90,13 +90,14 @@ class Basebootcamp:
return min(score * len(model_output) / short_threshold, score * len(ans_output) / ans_threshold, score * think_length / think_threshold)
# This for training Debug
if random.randint(1,1024) == 1:
print("=============DEBUG=============")
print("model_output:\n", model_output)
print("identity:\n", identity)
print("extract_solution:\n", extract_solution)
print("score:", score)
print("===============================")
# if random.randint(1,1024) == 1:
# print("=============DEBUG=============")
# print("model_output:\n", model_output)
# print("identity:\n", identity)
# print("extract_solution:\n", extract_solution)
# print("score:", score)
# print("===============================")
return score