mirror of
https://github.com/InternLM/InternBootcamp.git
synced 2026-04-24 17:05:00 +00:00
update to tech report version (#10)
* feat(run_eval): add checkpoint resume functionality and update example documentation; - update new bootcamp benchmark dataset * refactor(data_pipeline): optimize data generation pipeline; add multiple preset configurations for data generation * docs: update bootcamp list and add new scripts - Update Fulllist_InternBootcamp.md with new bootcamps and categories - Add new scripts to .gitignore: - examples/pipelines/filter_autogen_configs.py - examples/pipelines/quickgen_data_configs_from_eval_meta.py - Update dependencies in setup.py: - Add scipy and scikit-learn * refactor(internbootcamp): update bootcamp modules and improve error handling - Update import statements in __init__.py files - Add timestamp to target directory name in verl_data_preprocess.py - Improve error handling and scoring logic in bootcamp_judger.py - Remove unnecessary comments and update puzzle descriptions in multiple files
This commit is contained in:
parent
125a7818e0
commit
a8249acc18
2952 changed files with 105460 additions and 17649 deletions
|
|
@ -19,8 +19,8 @@ class bootcampJudger(BaseJudger):
|
|||
self,
|
||||
stop_word="<|im_end|>",
|
||||
format_score=0,
|
||||
format_penalty=True,
|
||||
short_penalty=True,
|
||||
format_penalty=False,
|
||||
short_penalty=False,
|
||||
short_threshold=128,
|
||||
|
||||
):
|
||||
|
|
@ -70,9 +70,11 @@ class bootcampJudger(BaseJudger):
|
|||
bootcamp_cls= getattr(internbootcamp, data_source[0].upper() + data_source[1:] + "bootcamp")
|
||||
try:
|
||||
score = bootcamp_cls.verify_score(response,identity,format_score=self.format_score,format_penalty=self.format_penalty,short_penalty=self.short_penalty,short_threshold=self.short_threshold)
|
||||
except:
|
||||
score = bootcamp_cls.verify_score(response,identity,format_score=self.format_score)
|
||||
return score
|
||||
return float(score)
|
||||
except Exception as e:
|
||||
print(f"Error in bootcampJudger {data_source}: {e}")
|
||||
return 0.0
|
||||
|
||||
# print(f"[Debug] Prompt: {prompt}")
|
||||
# print(f"[Debug]: score: {score}, response: {response}")
|
||||
# if type(score) == int:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue