mirror of
https://github.com/open-thought/reasoning-gym.git
synced 2026-04-19 12:58:07 +00:00
feat: Add LCM dataset generator for arithmetic reasoning tasks
This commit is contained in:
parent
2bc9319aa6
commit
8d369e6ced
4 changed files with 239 additions and 0 deletions
|
|
@ -10,6 +10,7 @@ The goal is to generate virtually infinite data with adjustable complexity.
|
||||||
- `ArithmeticDataset`: Generate arithmetic expressions with configurable complexity and operators (+, -, *)
|
- `ArithmeticDataset`: Generate arithmetic expressions with configurable complexity and operators (+, -, *)
|
||||||
- `ChainSum`: Generate addition/subtraction chains with configurable length and digit counts
|
- `ChainSum`: Generate addition/subtraction chains with configurable length and digit counts
|
||||||
- `GCDDataset`: Generate Greatest Common Divisor problems with configurable number of integers
|
- `GCDDataset`: Generate Greatest Common Divisor problems with configurable number of integers
|
||||||
|
- `LCMDataset`: Generate Least Common Multiple problems with configurable number of integers
|
||||||
- `LegCountingDataset`: Generate animal leg counting word problems with various animals
|
- `LegCountingDataset`: Generate animal leg counting word problems with various animals
|
||||||
- `PrimeFactorizationDataset`: Generate prime factorization tasks with configurable number ranges
|
- `PrimeFactorizationDataset`: Generate prime factorization tasks with configurable number ranges
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -9,6 +9,7 @@ Arithmetic tasks for training reasoning capabilities:
|
||||||
from .basic_arithmetic import ArithmeticDataset, ArithmeticDatasetConfig, arithmetic_dataset
|
from .basic_arithmetic import ArithmeticDataset, ArithmeticDatasetConfig, arithmetic_dataset
|
||||||
from .chain_sum import ChainSum, ChainSumConfig, chain_sum_dataset
|
from .chain_sum import ChainSum, ChainSumConfig, chain_sum_dataset
|
||||||
from .gcd import GCDConfig, GCDDataset, gcd_dataset
|
from .gcd import GCDConfig, GCDDataset, gcd_dataset
|
||||||
|
from .lcm import LCMConfig, LCMDataset, lcm_dataset
|
||||||
from .leg_counting import LegCountingConfig, LegCountingDataset, leg_counting_dataset
|
from .leg_counting import LegCountingConfig, LegCountingDataset, leg_counting_dataset
|
||||||
from .prime_factorization import PrimeFactorizationConfig, PrimeFactorizationDataset, prime_factorization_dataset
|
from .prime_factorization import PrimeFactorizationConfig, PrimeFactorizationDataset, prime_factorization_dataset
|
||||||
|
|
||||||
|
|
@ -22,6 +23,9 @@ __all__ = [
|
||||||
"GCDConfig",
|
"GCDConfig",
|
||||||
"GCDDataset",
|
"GCDDataset",
|
||||||
"gcd_dataset",
|
"gcd_dataset",
|
||||||
|
"LCMConfig",
|
||||||
|
"LCMDataset",
|
||||||
|
"lcm_dataset",
|
||||||
"LegCountingConfig",
|
"LegCountingConfig",
|
||||||
"LegCountingDataset",
|
"LegCountingDataset",
|
||||||
"leg_counting_dataset",
|
"leg_counting_dataset",
|
||||||
|
|
|
||||||
95
reasoning_gym/arithmetic/lcm.py
Normal file
95
reasoning_gym/arithmetic/lcm.py
Normal file
|
|
@ -0,0 +1,95 @@
|
||||||
|
"""Least Common Multiple (LCM) task generator"""
|
||||||
|
from dataclasses import dataclass
|
||||||
|
from random import Random
|
||||||
|
from typing import List, Optional, Tuple
|
||||||
|
from math import lcm
|
||||||
|
from functools import reduce
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class LCMConfig:
|
||||||
|
"""Configuration for LCM task generation"""
|
||||||
|
min_numbers: int = 2 # Minimum numbers to find LCM of
|
||||||
|
max_numbers: int = 2 # Maximum numbers to find LCM of
|
||||||
|
min_value: int = 1 # Minimum value for each number
|
||||||
|
max_value: int = 100 # Maximum value for each number (kept smaller than GCD default since LCM grows fast)
|
||||||
|
seed: Optional[int] = None
|
||||||
|
size: int = 500 # Virtual dataset size
|
||||||
|
|
||||||
|
def validate(self):
|
||||||
|
"""Validate configuration parameters"""
|
||||||
|
assert self.min_numbers >= 2, "min_numbers must be at least 2"
|
||||||
|
assert self.max_numbers >= self.min_numbers, "max_numbers must be >= min_numbers"
|
||||||
|
assert self.min_value >= 1, "min_value must be positive"
|
||||||
|
assert self.max_value > self.min_value, "max_value must be > min_value"
|
||||||
|
|
||||||
|
|
||||||
|
class LCMDataset:
|
||||||
|
"""Generates Least Common Multiple (LCM) tasks"""
|
||||||
|
|
||||||
|
def __init__(self, config: LCMConfig):
|
||||||
|
self.config = config
|
||||||
|
self.config.validate()
|
||||||
|
self.seed = config.seed if config.seed is not None else Random().randint(0, 2**32)
|
||||||
|
|
||||||
|
def __len__(self) -> int:
|
||||||
|
return self.config.size
|
||||||
|
|
||||||
|
def __iter__(self):
|
||||||
|
self._current_idx = 0
|
||||||
|
return self
|
||||||
|
|
||||||
|
def __next__(self):
|
||||||
|
if self._current_idx >= self.config.size:
|
||||||
|
raise StopIteration
|
||||||
|
item = self[self._current_idx]
|
||||||
|
self._current_idx += 1
|
||||||
|
return item
|
||||||
|
|
||||||
|
def _generate_numbers(self, rng: Random) -> List[int]:
|
||||||
|
"""Generate a list of random positive integers"""
|
||||||
|
num_count = rng.randint(self.config.min_numbers, self.config.max_numbers)
|
||||||
|
return [rng.randint(self.config.min_value, self.config.max_value)
|
||||||
|
for _ in range(num_count)]
|
||||||
|
|
||||||
|
def _calculate_lcm(self, numbers: List[int]) -> int:
|
||||||
|
"""Calculate the LCM of a list of numbers"""
|
||||||
|
return reduce(lcm, numbers)
|
||||||
|
|
||||||
|
def __getitem__(self, idx: int) -> dict:
|
||||||
|
"""Generate a single LCM task"""
|
||||||
|
rng = Random(self.seed + idx)
|
||||||
|
|
||||||
|
numbers = self._generate_numbers(rng)
|
||||||
|
result = self._calculate_lcm(numbers)
|
||||||
|
|
||||||
|
numbers_str = ", ".join(str(n) for n in numbers)
|
||||||
|
|
||||||
|
return {
|
||||||
|
"question": f"Find the Least Common Multiple (LCM) of these numbers: {numbers_str}",
|
||||||
|
"answer": str(result),
|
||||||
|
"metadata": {
|
||||||
|
"numbers": numbers,
|
||||||
|
"result": result
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def lcm_dataset(
|
||||||
|
min_numbers: int = 2,
|
||||||
|
max_numbers: int = 2,
|
||||||
|
min_value: int = 1,
|
||||||
|
max_value: int = 100,
|
||||||
|
seed: Optional[int] = None,
|
||||||
|
size: int = 500,
|
||||||
|
) -> LCMDataset:
|
||||||
|
"""Create a LCMDataset with the given configuration."""
|
||||||
|
config = LCMConfig(
|
||||||
|
min_numbers=min_numbers,
|
||||||
|
max_numbers=max_numbers,
|
||||||
|
min_value=min_value,
|
||||||
|
max_value=max_value,
|
||||||
|
seed=seed,
|
||||||
|
size=size,
|
||||||
|
)
|
||||||
|
return LCMDataset(config)
|
||||||
139
tests/test_lcm.py
Normal file
139
tests/test_lcm.py
Normal file
|
|
@ -0,0 +1,139 @@
|
||||||
|
import pytest
|
||||||
|
from math import lcm
|
||||||
|
from functools import reduce
|
||||||
|
from reasoning_gym.arithmetic import LCMDataset, LCMConfig
|
||||||
|
|
||||||
|
|
||||||
|
def test_lcm_config_validation():
|
||||||
|
"""Test that invalid configs raise appropriate errors"""
|
||||||
|
with pytest.raises(AssertionError):
|
||||||
|
config = LCMConfig(min_numbers=1) # Should be >= 2
|
||||||
|
config.validate()
|
||||||
|
|
||||||
|
with pytest.raises(AssertionError):
|
||||||
|
config = LCMConfig(min_numbers=3, max_numbers=2) # max should be >= min
|
||||||
|
config.validate()
|
||||||
|
|
||||||
|
with pytest.raises(AssertionError):
|
||||||
|
config = LCMConfig(min_value=0) # Should be positive
|
||||||
|
config.validate()
|
||||||
|
|
||||||
|
with pytest.raises(AssertionError):
|
||||||
|
config = LCMConfig(min_value=100, max_value=50) # max should be > min
|
||||||
|
config.validate()
|
||||||
|
|
||||||
|
|
||||||
|
def test_lcm_deterministic():
|
||||||
|
"""Test that dataset generates same items with same seed"""
|
||||||
|
config = LCMConfig(seed=42, size=10)
|
||||||
|
dataset1 = LCMDataset(config)
|
||||||
|
dataset2 = LCMDataset(config)
|
||||||
|
|
||||||
|
for i in range(len(dataset1)):
|
||||||
|
assert dataset1[i] == dataset2[i]
|
||||||
|
|
||||||
|
|
||||||
|
def test_lcm_items():
|
||||||
|
"""Test basic properties of generated items"""
|
||||||
|
config = LCMConfig(
|
||||||
|
min_numbers=2,
|
||||||
|
max_numbers=4,
|
||||||
|
min_value=1,
|
||||||
|
max_value=20, # Keep small for testing
|
||||||
|
size=50,
|
||||||
|
seed=42
|
||||||
|
)
|
||||||
|
dataset = LCMDataset(config)
|
||||||
|
|
||||||
|
for i in range(len(dataset)):
|
||||||
|
item = dataset[i]
|
||||||
|
assert isinstance(item, dict)
|
||||||
|
assert "question" in item
|
||||||
|
assert "answer" in item
|
||||||
|
assert "metadata" in item
|
||||||
|
|
||||||
|
# Verify the numbers and result are in metadata
|
||||||
|
metadata = item["metadata"]
|
||||||
|
assert "numbers" in metadata
|
||||||
|
assert "result" in metadata
|
||||||
|
|
||||||
|
# Verify the numbers are within configured range
|
||||||
|
numbers = metadata["numbers"]
|
||||||
|
assert all(config.min_value <= n <= config.max_value for n in numbers)
|
||||||
|
assert config.min_numbers <= len(numbers) <= config.max_numbers
|
||||||
|
|
||||||
|
# Verify the LCM calculation is correct
|
||||||
|
result = metadata["result"]
|
||||||
|
assert str(result) == item["answer"]
|
||||||
|
assert result == reduce(lcm, numbers)
|
||||||
|
|
||||||
|
|
||||||
|
def test_lcm_number_ranges():
|
||||||
|
"""Test that generated numbers respect value constraints"""
|
||||||
|
config = LCMConfig(
|
||||||
|
min_numbers=2,
|
||||||
|
max_numbers=2,
|
||||||
|
min_value=5,
|
||||||
|
max_value=15,
|
||||||
|
size=20,
|
||||||
|
seed=42
|
||||||
|
)
|
||||||
|
dataset = LCMDataset(config)
|
||||||
|
|
||||||
|
for i in range(len(dataset)):
|
||||||
|
item = dataset[i]
|
||||||
|
numbers = item["metadata"]["numbers"]
|
||||||
|
assert all(5 <= n <= 15 for n in numbers)
|
||||||
|
|
||||||
|
|
||||||
|
def test_lcm_iteration():
|
||||||
|
"""Test that iteration works correctly"""
|
||||||
|
config = LCMConfig(size=5, seed=42)
|
||||||
|
dataset = LCMDataset(config)
|
||||||
|
|
||||||
|
# Test manual iteration
|
||||||
|
items = []
|
||||||
|
for item in dataset:
|
||||||
|
items.append(item)
|
||||||
|
assert len(items) == config.size
|
||||||
|
|
||||||
|
# Test list conversion
|
||||||
|
items = list(dataset)
|
||||||
|
assert len(items) == config.size
|
||||||
|
|
||||||
|
# Test multiple iterations yield same results
|
||||||
|
first_items = list(dataset)
|
||||||
|
second_items = list(dataset)
|
||||||
|
assert first_items == second_items
|
||||||
|
|
||||||
|
|
||||||
|
def test_lcm_special_cases():
|
||||||
|
"""Test some special LCM cases"""
|
||||||
|
config = LCMConfig(
|
||||||
|
min_numbers=2,
|
||||||
|
max_numbers=2,
|
||||||
|
min_value=1,
|
||||||
|
max_value=20,
|
||||||
|
size=100,
|
||||||
|
seed=42
|
||||||
|
)
|
||||||
|
dataset = LCMDataset(config)
|
||||||
|
|
||||||
|
# Track if we see some interesting LCM cases
|
||||||
|
seen_equal_to_product = False # When numbers are coprime
|
||||||
|
seen_less_than_product = False # When numbers share factors
|
||||||
|
|
||||||
|
for i in range(len(dataset)):
|
||||||
|
item = dataset[i]
|
||||||
|
numbers = item["metadata"]["numbers"]
|
||||||
|
result = int(item["answer"])
|
||||||
|
product = reduce(lambda x, y: x * y, numbers)
|
||||||
|
|
||||||
|
if result == product:
|
||||||
|
seen_equal_to_product = True
|
||||||
|
if result < product:
|
||||||
|
seen_less_than_product = True
|
||||||
|
|
||||||
|
# With enough samples, we should see both cases
|
||||||
|
assert seen_equal_to_product, "Expected to see some coprime numbers (LCM = product)"
|
||||||
|
assert seen_less_than_product, "Expected to see some numbers with common factors (LCM < product)"
|
||||||
Loading…
Add table
Add a link
Reference in a new issue