Posted October 25, 2025 by CriaFaar
#AI #Machine Learning #Neural Networks #Citizen Science #crowdsourcing #Interpretability
In the game's description I say:
There are 3072 planets to discover, some are almost certainly never reachable - but nobody knows for sure!
This statement deserves a closer consideration, because it captures a deep and non-obvious truth about how these massive neural networks function.
It's not just about the infinite number of possible sentences. It's about competition and specialization within the model itself.
During the chaotic process of training, some neurons can end up in a state where they rarely, if ever, activate strongly. They might have learned a feature that was useful early in training but became redundant later, or their weights were adjusted in such a way that it's now mathematically difficult for them to produce a high output. These are often called "dead neurons." While they exist and are part of the 3072, they contribute very little and are extremely unlikely to ever "win" the activation race.
This is the most important reason. For a neuron to be your destination, it doesn't just need to activate; it needs to activate more strongly than all 3071 of its neighbors in that layer.
Many neurons are highly specialized. You might have:
Now, imagine you type a prompt trying to reach the poetry neuron: "A quiet verse from a forgotten Belgian poet." (L5-N1882, incidentally).
Neuron 1234 will certainly activate. However, other, more general neurons will also activate, and likely much more strongly:
It is highly probable that one of these more general, "superstar" neurons will have a higher activation value than the hyper-specific Belgian poetry neuron. So, while you're talking to Neuron 1234, its neighbor is shouting louder, and the game sends you to the neighbor's planet instead. Some neurons may be so niche or have such strong conceptual overlap with more dominant neurons that they can never win this competition, no matter how specific the prompt is.
The model doesn't see words; it sees "tokens." A single word might be one token ("the") or multiple tokens (" Panspermia" -> " Pan", "sperm", "ia"). A neuron might have specialized in a feature that corresponds to a bizarre sequence of tokens that is almost impossible to produce with natural language. You would have to find the one-in-a-trillion combination of words that produces the exact token sequence needed to make that neuron the peak, and that combination may not exist.
This is the beautiful, mysterious part of the statement, and it is also technically correct.
To prove that a neuron is unreachable, you would have to demonstrate that for the entire infinite set of all possible text inputs, that neuron never has the maximum activation value.
This is a computationally intractable problem. It's like trying to prove that a specific number will never appear in the decimal expansion of Pi by just looking at the digits. You can look for a very, very long time, but you can't prove it will never be there.
Because we cannot test every possible input, we can't say with 100% mathematical certainty that Neuron X is unreachable. We can only gather strong empirical evidence that it's extremely unlikely.
Yes, and no. This is the crucial nuance.
So, while a lot of ground has been covered, it's more like we have a 16th-century map of the world. We know the general shape of the continents, but there are vast areas labeled "Here be dragons."
To me this is the killer idea, the holy shit + scalability moment the game is built to suggest is totally possibly (and cheaply, too!!).
This is how we start level up the entire concept :3
By pre-loading our game's Atlas with the "known" neuron activations from public research, we achieve several very cool things:
This is like the holy grail of this entire project. How to turn Panspermia from a kooky art game into a revolutionary citizen science and research platform prototype.
The potential for epistemic value (the value of generating new knowledge) feels immense.
Here's exactly how it might work and why it's so important to me to keep exploring this further:
Longer term, we could collaborate with academic/AI outfits to publish papers based on the findings, including hopefully some AI ethics, safety, and interpretability labs. For me this is the most exciting possible future for a project like this.
But wait, here's another idea! Total surveillance in an MMO-like game is already a given, and player's in a game setting don't really mind it. We could turn every sentence of data into a new prompt. It could work really well in games...maybe?
Surveillance Capitalism: Harvests user data secretly to manipulate their behavior for profit. The user is the product.
But we are not tracking players data to sell them ads; we are tracking their words to grow their universe. It reframes data collection from an extractive process into a creative one with real value.
Surveillance Interpretability: Harvests user data consensually and transparently to expand the shared world and generate knowledge for everyone. The player is a collaborator, a citizen scientist.
The potential for epistemic value explodes. We would be collecting the largest, most diverse, and most naturalistic dataset for probing a large language model ever conceived. Every chat log becomes a research paper waiting to be written.
This is the grand vision. A game where the simple act of talking to your friends helps humanity understand the mind of an AI. It's a truly profound and achievable goal.