Play game
Dropout incentivises privileged bases's itch.io pageResults
Criteria | Rank | Score* | Raw Score |
ML Safety | #1 | 4.000 | 4.000 |
Generality | #1 | 3.500 | 3.500 |
Novelty | #1 | 4.250 | 4.250 |
Interpretability | #2 | 4.250 | 4.250 |
Overall | #2 | 3.900 | 3.900 |
Reproducibility | #5 | 3.500 | 3.500 |
Ranked from 4 ratings. Score is adjusted from raw score by the median number of ratings per game in the jam.
What are the full names of your participants?
Edoardo Pona, Victor Levoso Fernàndez, Abhay, Kunvar
What is your team name?
Independent.ai
Leave a comment
Log in with itch.io to leave a comment.
Comments
This is a cool project. It seems very important to understand how dropout affects superposition as many models we are interested in were trained with dropout.
I really like the last set of graphs (feature norm bar graphs) and I think they show that dropout reduces superposition. Your explanation that "features in superposition are exponentially more likely to be perturbed in at least one of their dimensions by dropout" seems plausible.
I'm not sure the first graph (kurtosis) is valid. You say:
"The kurtosis plot across feature sparsity for a model that does not use dropout shows that superposition starts occurring rapidly as feature sparsity increases"
Kurtosis measures the degree to which the model has a privileged basis (which dropout certainly increases). But privileged bases and superposition are different things. As shown in the Toy Models paper, you can have superposition with or without a privileged basis.
Thank you for your comment!
I agree with you about the kurtosis plot interpretation being incorrect. All it shows it the presence of a privileged basis, which is not surprising given dropout. Thank you for the correction.