mirror of
https://github.com/open-thought/reasoning-gym.git
synced 2026-04-29 17:35:16 +00:00
added curriculum
This commit is contained in:
parent
eb69916c1b
commit
6fa76f11b5
19 changed files with 1104 additions and 450 deletions
|
|
@ -1,5 +1,6 @@
|
|||
from typing import Optional
|
||||
from typing import Literal, Optional
|
||||
|
||||
import numpy as np
|
||||
import verl.utils.torch_functional as verl_F
|
||||
from torch.utils.data import Dataset
|
||||
from transformers import PreTrainedTokenizer
|
||||
|
|
@ -25,6 +26,11 @@ class ReasoningGymDataset(Dataset):
|
|||
procedural_dataset is None or experiment is None
|
||||
), "Only one of `procedural_dataset` or `experiment` may be provided"
|
||||
|
||||
assert procedural_dataset or experiment, "One of `procedural_dataset` or `experiment` must be provided"
|
||||
assert (
|
||||
procedural_dataset is None or experiment is None
|
||||
), "Only one of `procedural_dataset` or `experiment` may be provided"
|
||||
|
||||
self.tokenizer = tokenizer
|
||||
self.data = procedural_dataset or experiment.composite
|
||||
self.experiment = experiment
|
||||
|
|
@ -67,10 +73,39 @@ class ReasoningGymDataset(Dataset):
|
|||
row_dict["index"] = index
|
||||
return row_dict
|
||||
|
||||
def update_experiment_difficulty(self, dataset_name: str, method: Literal["increment", "decrement"]):
|
||||
"""Update the difficulty of the underlying dataset."""
|
||||
if self.experiment is None:
|
||||
raise ValueError("Cannot update difficulty: dataset is not a CurriculumExperiment")
|
||||
if method not in ["increment", "decrement"]:
|
||||
raise ValueError("Invalid method: must be 'increment' or 'decrement'")
|
||||
self.experiment.score_board.clear()
|
||||
self.experiment.update_difficulty(dataset_name, method)
|
||||
self.data = self.experiment.composite
|
||||
return True
|
||||
|
||||
def aggregate(self, last_n: Optional[int] = None):
|
||||
"""Aggregate scores from the underlying experiment"""
|
||||
if self.experiment is None:
|
||||
raise ValueError("Cannot aggregate scores: dataset is not a CurriculumExperiment")
|
||||
|
||||
results = self.experiment.score_board.aggregate(last_n=last_n)
|
||||
output_results = {}
|
||||
|
||||
for key, value in results.items():
|
||||
output_results[key] = {}
|
||||
scores = value.scores
|
||||
first_key = list(scores.keys())[0]
|
||||
output_results[key]["results"] = np.mean(scores[first_key])
|
||||
output_results[key]["total_samples"] = value.total_scores
|
||||
|
||||
return output_results
|
||||
|
||||
|
||||
def make_dataset(
|
||||
tokenizer,
|
||||
data_source: Experiment | ProceduralDataset,
|
||||
dataset_name: str,
|
||||
developer_prompt: str,
|
||||
) -> ReasoningGymDataset:
|
||||
"""
|
||||
|
|
@ -78,10 +113,12 @@ def make_dataset(
|
|||
"""
|
||||
kwargs = {
|
||||
"tokenizer": tokenizer,
|
||||
# "dataset_name": dataset_name,
|
||||
"developer_prompt": developer_prompt,
|
||||
}
|
||||
if isinstance(data_source, Experiment):
|
||||
kwargs["experiment"] = data_source
|
||||
else:
|
||||
kwargs["procedural_dataset"] = data_source
|
||||
print(type(data_source))
|
||||
return ReasoningGymDataset(**kwargs)
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue