reasoning-gym/training/utils/datasets.py
Oliver Stanley eb69916c1b
initial verl training codebase (#389)
* fixes for latest verl
* composite dataset training experiment
* use stateful dataloaders to match verl changes
* training readme
* add formatting reward
* length reward impl
* standalone reasoning_gym config section
* curriculum learning, new length reward, more config
2025-03-20 15:04:57 +00:00

87 lines
3.1 KiB
Python

from typing import Optional
import verl.utils.torch_functional as verl_F
from torch.utils.data import Dataset
from transformers import PreTrainedTokenizer
from verl.utils.model import compute_position_id_with_mask
from reasoning_gym.coaching.experiment import Experiment
from reasoning_gym.dataset import ProceduralDataset
class ReasoningGymDataset(Dataset):
def __init__(
self,
tokenizer: PreTrainedTokenizer,
procedural_dataset: Optional[ProceduralDataset] = None,
experiment: Optional[Experiment] = None,
developer_prompt: Optional[str] = None,
developer_role: str = "system",
max_prompt_length: int = 2048,
truncation: str = "error", ## ['left', 'right', 'error']
):
assert procedural_dataset or experiment, "One of `procedural_dataset` or `experiment` must be provided"
assert (
procedural_dataset is None or experiment is None
), "Only one of `procedural_dataset` or `experiment` may be provided"
self.tokenizer = tokenizer
self.data = procedural_dataset or experiment.composite
self.experiment = experiment
self.developer_prompt = developer_prompt
self.developer_role = developer_role
self.max_prompt_length = max_prompt_length
self.truncation = truncation
def __len__(self) -> int:
return len(self.data)
def __getitem__(self, index):
row_dict = self.data[index].copy()
q = row_dict["question"]
chat = []
if self.developer_prompt is not None:
chat.append({"role": self.developer_role, "content": self.developer_prompt})
chat.append({"role": "user", "content": q})
prompt = self.tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True)
input_ids, attention_mask = verl_F.tokenize_and_postprocess_data(
prompt=prompt,
tokenizer=self.tokenizer,
max_length=self.max_prompt_length,
pad_token_id=self.tokenizer.pad_token_id,
left_pad=True,
truncation=self.truncation,
)
position_ids = compute_position_id_with_mask(attention_mask)
row_dict["data_source"] = "reasoning_gym"
row_dict["input_ids"] = input_ids[0]
row_dict["attention_mask"] = attention_mask[0]
row_dict["position_ids"] = position_ids[0]
row_dict["raw_prompt_ids"] = self.tokenizer.encode(prompt, add_special_tokens=False)
row_dict["raw_prompt"] = chat
row_dict["index"] = index
return row_dict
def make_dataset(
tokenizer,
data_source: Experiment | ProceduralDataset,
developer_prompt: str,
) -> ReasoningGymDataset:
"""
Create ReasoningGymDataset object using either a ProceduralDataset or Experiment as the underlying data source.
"""
kwargs = {
"tokenizer": tokenizer,
"developer_prompt": developer_prompt,
}
if isinstance(data_source, Experiment):
kwargs["experiment"] = data_source
else:
kwargs["procedural_dataset"] = data_source
return ReasoningGymDataset(**kwargs)