include ranges rather than sampled values in difficulty metadata dicts (#387)

* update difficulty metadata for logic datasets

* update difficulty metadata for graph datasets

* update difficulty metadata for geometry datasets

* update difficulty metadata for games datasets

* update difficulty metadata for cognition datasets

* update difficulty metadata for arithmetic datasets

* update difficulty metadata for arc datasets

* update difficulty metadata for algorithmic datasets

* update difficulty metadata for algebra datasets

* use tuples

* update tests

* update tests
This commit is contained in:
Oliver Stanley 2025-03-20 09:27:03 +00:00 committed by GitHub
parent b69c35818a
commit 7475a20700
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
80 changed files with 304 additions and 126 deletions

View file

@ -387,7 +387,7 @@ class CircuitLogicDataset(ProceduralDataset):
"final_gate": final_gate_name,
"inputs": inputs_list,
"difficulty": {
"terms": num_terms,
"terms": (self.config.min_terms, self.config.max_terms),
"inputs": (self.config.min_inputs, self.config.max_inputs),
},
},

View file

@ -221,8 +221,8 @@ class PropositionalLogicDataset(ProceduralDataset):
"complexity": self._measure_complexity(conclusion),
"example_answer": str(conclusion),
"difficulty": {
"vars": num_vars,
"statements": num_statements,
"vars": (self.config.min_vars, self.config.max_vars),
"statements": (self.config.min_statements, self.config.max_statements),
"complexity": (self.config.min_complexity, self.config.max_complexity),
},
},

View file

@ -346,7 +346,9 @@ class SelfReferenceDataset(ProceduralDataset):
return {
"question": puzz_s,
"answer": answer,
"metadata": {"difficulty": difficulty},
"metadata": {
"difficulty": {"difficulty": difficulty},
},
}
def score_answer(self, answer: Optional[str], entry: dict[str, Any]) -> float: