* init
* fix tests
* unify codeio
* filtered for libraries not present in reasoning-gym
* fix more bounds
* puzzle24
* knight swap curriculum
* fix number sorting
* fix attributes
* add validation of config in creation of dataset
* dry run for instantiating and validating the datasets
* remove unused imports
* fix curriculum tests to reference newly updated attribute names
This is a variant of the Game of Life task, which rather than trying to test the algorithmic simulation, tests the ability of the model to do explanatory reasoning of the board. The idea is that a model with good explanatory reasoning will be able to see that a game will not halt without simulating it into the future.
The task presents a GoL board, and the model is asked to predict if the board will halt (die, all cells zero) after n steps. Sometimes, the board will be made up of 'oscillators', isolated structures which never die. Othertimes, it is filled with non-oscillators, structures which will always die after a few steps. The model should deduce which case the presented board is.