mirror of
https://github.com/open-thought/reasoning-gym.git
synced 2026-04-19 12:58:07 +00:00
188 lines
6.7 KiB
Python
188 lines
6.7 KiB
Python
import pytest
|
|
|
|
from reasoning_gym.games import FutoshikiConfig, FutoshikiDataset
|
|
|
|
|
|
def test_futoshiki_config_validation():
|
|
"""Test that invalid configs raise appropriate errors"""
|
|
with pytest.raises(AssertionError):
|
|
config = FutoshikiConfig(board_size=3) # Too small
|
|
config.validate()
|
|
|
|
with pytest.raises(AssertionError):
|
|
config = FutoshikiConfig(board_size=10) # Too large
|
|
config.validate()
|
|
|
|
with pytest.raises(AssertionError):
|
|
config = FutoshikiConfig(difficulty=-1) # Invalid difficulty
|
|
config.validate()
|
|
|
|
with pytest.raises(AssertionError):
|
|
config = FutoshikiConfig(difficulty=4) # Invalid difficulty
|
|
config.validate()
|
|
|
|
|
|
def test_futoshiki_deterministic():
|
|
"""Test that dataset generates same puzzles with same seed"""
|
|
config = FutoshikiConfig(seed=42, size=10)
|
|
dataset1 = FutoshikiDataset(config)
|
|
dataset2 = FutoshikiDataset(config)
|
|
|
|
for i in range(len(dataset1)):
|
|
assert dataset1[i] == dataset2[i]
|
|
|
|
|
|
def test_futoshiki_items():
|
|
"""Test basic properties of generated items"""
|
|
config = FutoshikiConfig(board_size=4, difficulty=1, size=10, seed=42)
|
|
dataset = FutoshikiDataset(config)
|
|
|
|
for i in range(len(dataset)):
|
|
item = dataset[i]
|
|
assert isinstance(item, dict)
|
|
assert "question" in item
|
|
assert "answer" in item
|
|
assert "metadata" in item
|
|
|
|
# Verify metadata contents
|
|
metadata = item["metadata"]
|
|
assert "puzzle" in metadata
|
|
assert "solution" in metadata
|
|
assert "constraints" in metadata
|
|
assert "board_size" in metadata
|
|
assert "difficulty" in metadata
|
|
|
|
# Verify board dimensions
|
|
puzzle = metadata["puzzle"]
|
|
solution = metadata["solution"]
|
|
assert len(puzzle) == config.board_size
|
|
assert len(solution) == config.board_size
|
|
for row in puzzle:
|
|
assert len(row) == config.board_size
|
|
for row in solution:
|
|
assert len(row) == config.board_size
|
|
|
|
# Verify constraints format
|
|
constraints = metadata["constraints"]
|
|
for ((r1, c1), (r2, c2)), rel in constraints.items():
|
|
assert 0 <= r1 < config.board_size
|
|
assert 0 <= c1 < config.board_size
|
|
assert 0 <= r2 < config.board_size
|
|
assert 0 <= c2 < config.board_size
|
|
assert rel in ("<", ">")
|
|
|
|
|
|
def test_futoshiki_solution_validity():
|
|
"""Test that solutions are valid according to Futoshiki rules"""
|
|
config = FutoshikiConfig(board_size=4, difficulty=1, size=10, seed=42)
|
|
dataset = FutoshikiDataset(config)
|
|
|
|
def is_valid_solution(solution, board_size, constraints):
|
|
# Check rows
|
|
for row in solution:
|
|
if sorted(row) != list(range(1, board_size + 1)):
|
|
return False
|
|
|
|
# Check columns
|
|
for col in range(board_size):
|
|
column = [solution[row][col] for row in range(board_size)]
|
|
if sorted(column) != list(range(1, board_size + 1)):
|
|
return False
|
|
|
|
# Check constraints
|
|
for ((r1, c1), (r2, c2)), rel in constraints.items():
|
|
v1, v2 = solution[r1][c1], solution[r2][c2]
|
|
if rel == "<" and not (v1 < v2):
|
|
return False
|
|
if rel == ">" and not (v1 > v2):
|
|
return False
|
|
|
|
return True
|
|
|
|
for i in range(len(dataset)):
|
|
item = dataset[i]
|
|
metadata = item["metadata"]
|
|
solution = metadata["solution"]
|
|
constraints = metadata["constraints"]
|
|
|
|
assert is_valid_solution(solution, config.board_size, constraints)
|
|
|
|
|
|
def test_futoshiki_puzzle_solvability():
|
|
"""Test that generated puzzles are solvable and have unique solutions"""
|
|
config = FutoshikiConfig(board_size=4, difficulty=1, size=5, seed=42)
|
|
dataset = FutoshikiDataset(config)
|
|
|
|
for i in range(len(dataset)):
|
|
item = dataset[i]
|
|
metadata = item["metadata"]
|
|
puzzle = metadata["puzzle"]
|
|
constraints = metadata["constraints"]
|
|
|
|
# Verify puzzle has exactly one solution
|
|
assert dataset.count_solutions(puzzle, constraints, limit=2) == 1
|
|
|
|
|
|
def test_futoshiki_difficulty_levels():
|
|
"""Test that different difficulty levels affect puzzle complexity"""
|
|
size = 5
|
|
board_size = 4
|
|
seeds = [42, 43, 44] # Test multiple seeds for robustness
|
|
|
|
def count_clues(puzzle):
|
|
return sum(cell != 0 for row in puzzle for cell in row)
|
|
|
|
def count_constraints(constraints):
|
|
return len(constraints)
|
|
|
|
for seed in seeds:
|
|
clues_by_difficulty = []
|
|
constraints_by_difficulty = []
|
|
|
|
for difficulty in range(4): # 0 to 3
|
|
config = FutoshikiConfig(board_size=board_size, difficulty=difficulty, size=size, seed=seed)
|
|
dataset = FutoshikiDataset(config)
|
|
|
|
avg_clues = sum(count_clues(item["metadata"]["puzzle"]) for item in dataset) / size
|
|
avg_constraints = sum(count_constraints(item["metadata"]["constraints"]) for item in dataset) / size
|
|
|
|
clues_by_difficulty.append(avg_clues)
|
|
constraints_by_difficulty.append(avg_constraints)
|
|
|
|
# Higher difficulty should generally mean fewer clues and/or more constraints
|
|
assert all(clues_by_difficulty[i] >= clues_by_difficulty[i + 1] for i in range(len(clues_by_difficulty) - 1))
|
|
assert all(
|
|
constraints_by_difficulty[i] <= constraints_by_difficulty[i + 1]
|
|
for i in range(len(constraints_by_difficulty) - 1)
|
|
)
|
|
|
|
|
|
def test_futoshiki_answer_scoring():
|
|
"""Test the answer scoring mechanism"""
|
|
config = FutoshikiConfig(board_size=4, difficulty=0, size=5, seed=42)
|
|
dataset = FutoshikiDataset(config)
|
|
|
|
for item in dataset:
|
|
# Correct answer should score 1.0
|
|
assert dataset.score_answer(item["answer"], item) == 1.0
|
|
|
|
# Wrong answer should score lower
|
|
wrong_answer = item["answer"].replace("1", "2")
|
|
assert dataset.score_answer(wrong_answer, item) < 1.0
|
|
|
|
# None or empty answer should score 0.0
|
|
assert dataset.score_answer(None, item) == 0.0
|
|
assert dataset.score_answer("", item) == 0.0
|
|
|
|
answer = item["answer"]
|
|
white_space_mismatch = answer.replace(" ", " ")
|
|
assert dataset.score_answer(white_space_mismatch, item) == 0.9
|
|
|
|
anwser_with_additional_text = "This is an anwser " + answer + "\nwith surrounding text."
|
|
assert 0 < dataset.score_answer(anwser_with_additional_text, item) < 0.9
|
|
|
|
partially_correct = anwser_with_additional_text.replace("1", "2")
|
|
assert dataset.score_answer(partially_correct, item) > 0.1
|
|
|
|
bad_answer = "\n".join(anwser_with_additional_text.split("\n")[::-1])
|
|
assert dataset.score_answer(bad_answer, item) < 0.1
|