reasoning-gym/tests/test_arc_1d.py
Zafir Stojanovski dced3bfc45
fix(curriculum): Make boundaries in curriculum more sensible (#407)
* init

* fix tests

* unify codeio

* filtered for libraries not present in reasoning-gym

* fix more bounds

* puzzle24

* knight swap curriculum

* fix number sorting

* fix attributes

* add validation of config in creation of dataset

* dry run for instantiating and validating the datasets

* remove unused imports

* fix curriculum tests to reference newly updated attribute names
2025-04-04 20:24:14 +02:00

186 lines
6.3 KiB
Python

from random import Random
import pytest
from reasoning_gym.arc import Arc1DConfig, Arc1DCurriculum, Arc1DDataset
def test_arc_1d_config_validation():
"""Test that invalid configs raise appropriate errors"""
with pytest.raises(AssertionError):
config = Arc1DConfig(min_size=0)
config.validate()
with pytest.raises(AssertionError):
config = Arc1DConfig(min_size=30, max_size=20)
config.validate()
with pytest.raises(AssertionError):
config = Arc1DConfig(num_train=0)
config.validate()
def test_arc_1d_deterministic():
"""Test that dataset generates same items with same seed"""
config = Arc1DConfig(seed=42, size=10)
dataset1 = Arc1DDataset(config)
dataset2 = Arc1DDataset(config)
for i in range(len(dataset1)):
assert dataset1[i] == dataset2[i]
def test_arc_1d_items():
"""Test basic properties of generated items"""
config = Arc1DConfig(min_size=10, max_size=15, num_train=2, size=50, seed=42)
dataset = Arc1DDataset(config)
for i in range(len(dataset)):
item = dataset[i]
assert isinstance(item, dict)
assert "question" in item
assert "answer" in item
assert "metadata" in item
assert "difficulty" in item["metadata"]
# Check metadata contents
metadata = item["metadata"]
assert "task_name" in metadata
assert "size" in metadata
assert "train_examples" in metadata
assert "test_example" in metadata
# Verify size constraints
assert config.min_size <= metadata["size"] <= config.max_size
# Check training examples
train_examples = metadata["train_examples"]
assert len(train_examples) == config.num_train
for example in train_examples:
assert "input" in example
assert "output" in example
assert len(example["input"]) == metadata["size"]
assert len(example["output"]) == metadata["size"]
# Check test example
test_example = metadata["test_example"]
assert "input" in test_example
assert "output" in test_example
assert len(test_example["input"]) == metadata["size"]
assert len(test_example["output"]) == metadata["size"]
def test_arc_1d_iteration():
"""Test that iteration respects dataset size"""
config = Arc1DConfig(size=100, seed=42) # Small size for testing
dataset = Arc1DDataset(config)
# Test manual iteration
items = []
for item in dataset:
items.append(item)
assert len(items) == config.size, "Iterator should yield exactly size items"
# Test list conversion
items = list(dataset)
assert len(items) == config.size, "Iterator should yield exactly size items"
# Test multiple iterations
first_items = list(dataset)
second_items = list(dataset)
assert first_items == second_items, "Multiple iterations should yield same items"
def test_arc_1d_scoring():
"""Test answer scoring logic"""
config = Arc1DConfig(size=1, seed=42)
dataset = Arc1DDataset(config)
entry = dataset[0]
# Test exact match
assert dataset.score_answer(entry["answer"], entry) == 1.0
# Test partial match (answer contained within response)
assert dataset.score_answer(f"The answer is: {entry['answer']}", entry) > 0.5
# Test incorrect answer
assert dataset.score_answer("wrong answer", entry) == 0.0
# Test None answer
assert dataset.score_answer(None, entry) == 0.0
@pytest.mark.parametrize("board_size", [8, 9, 10, 12, 15, 20])
def test_arc_1d_sizes(board_size: int):
config = Arc1DConfig(size=1000, seed=42 + board_size, min_size=board_size, max_size=board_size)
dataset = Arc1DDataset(config)
for entry in dataset:
assert len(entry["metadata"]["test_example"]["input"]) == board_size
assert len(entry["metadata"]["test_example"]["output"]) == board_size
assert dataset.score_answer(entry["answer"], entry) == 1.0
@pytest.mark.parametrize("min_size,max_size", [(8, 10), (9, 13), (10, 12), (12, 20)])
def test_arc_1d_size_ranges(min_size: int, max_size: int):
config = Arc1DConfig(size=1000, seed=42, min_size=min_size, max_size=max_size)
dataset = Arc1DDataset(config)
for entry in dataset:
assert min_size <= len(entry["metadata"]["test_example"]["input"]) <= max_size
assert min_size <= len(entry["metadata"]["test_example"]["output"]) <= max_size
assert dataset.score_answer(entry["answer"], entry) == 1.0
def test_arc_1d_generate_all_tasks():
config = Arc1DConfig(size=100, seed=17, min_size=8, max_size=10)
dataset = Arc1DDataset(config)
tasks = dataset.ARC_1D_TASKS
rng = Random(999)
for task_name, (generator_fn, args) in tasks.items():
for j in range(3):
for i in range(20):
x = generator_fn(rng=rng, size=10, **args)
if x is not None:
break
assert i < 20
print(task_name, j, i, x)
def test_arc_1d_curriculum():
"""Test the curriculum for complex arithmetic."""
curriculum = Arc1DCurriculum()
base_value = {"size": 150, "seed": 1}
base_cfg: Arc1DCurriculum = curriculum.generate_configuration(base_value)
assert base_cfg.seed == 1
assert base_cfg.size == 150
assert base_cfg.min_size == 10
assert base_cfg.max_size == 10
# Test and validate increase in levels
curriculum.increment_attr_level("size")
increased_cfg: Arc1DCurriculum = curriculum.generate_configuration(base_value)
assert increased_cfg.min_size == 10
assert increased_cfg.max_size == 25
# Test and validate decrease in levels
curriculum.decrement_attr_level("size")
decreased_cfg: Arc1DCurriculum = curriculum.generate_configuration(base_value)
assert decreased_cfg.min_size == 10
assert decreased_cfg.max_size == 10
# Test upper bound boundary condition
for _ in range(10):
curriculum.increment_attr_level("size")
upper_bound_cfg: Arc1DCurriculum = curriculum.generate_configuration(base_value)
assert upper_bound_cfg.min_size == 10
assert upper_bound_cfg.max_size == 100
# Test lower bound boundary condition
for _ in range(10):
curriculum.decrement_attr_level("size")
lower_bound_cfg: Arc1DCurriculum = curriculum.generate_configuration(base_value)
assert lower_bound_cfg.min_size == 10
assert lower_bound_cfg.max_size == 10