This post was rejected for the following reason(s):
No LLM generated, heavily assisted/co-written, or otherwise reliant work. LessWrong has recently been inundated with new users submitting work where much of the content is the output of LLM(s). This work by-and-large does not meet our standards, and is rejected. This includes dialogs with LLMs that claim to demonstrate various properties about them, posts introducing some new concept and terminology that explains how LLMs work, often centered around recursiveness, emergence, sentience, consciousness, etc. Our LLM-generated content policy can be viewed here.
Insufficient Quality for AI Content. There’ve been a lot of new users coming to LessWrong recently interested in AI. To keep the site’s quality high and ensure stuff posted is interesting to the site’s users, we’re currently only accepting posts that meet a pretty high bar.
If you want to try again, I recommend writing something short and to the point, focusing on your strongest argument, rather than a long, comprehensive essay. (This is fairly different from common academic norms.) We get lots of AI essays/papers every day and sadly most of them don't make very clear arguments, and we don't have time to review them all thoroughly.
We look for good reasoning, making a new and interesting point, bringing new evidence, and/or building upon prior discussion. If you were rejected for this reason, possibly a good thing to do is read more existing material. The AI Intro Material wiki-tag is a good place, for example.
This short post introduces EvoNet, a prototype of a novel neural architecture that maintains a persistent internal state, routes activations bidirectionally, groups nodes into clusters, and remains GPU-friendly through iterative updates. The core question: could such an architecture transport and refine information in a way that can compete with conventional ANNs?
A New Structure: Persistent State and Flowing Activations
Unlike typical ANNs (which clear their activations after each forward pass), EvoNet maintains a persistent internal state. Signals aren't just pushed forward and discarded — they continue to circulate and evolve with each iteration. This is more akin to how biological brains maintain working memory and adapt responses over time.
Information-Flow Diagrams: Conventional ANN vs EvoNet
Conventional ANNEvoNet
In conventional networks, information moves in a single sweep. In EvoNet, activations loop back and cross paths, letting the network build a richer internal context. Whether this is truly viable remains an open question. Can such recursive flows reliably transport meaningful signals, or do they collapse into noise or a static state?
Nodes and Clusters: Scalable Building Blocks
EvoNet's micro-units are nodes, similar to ANN neurons but with local state updates every iteration. This lets signals ripple like waves, not just be processed once. Nodes group into clusters, forming natural units for scaling and specialization. This could later support dynamic growth by inserting nodes or clusters precisely where informational "strain" emerges.
These ideas for network growth will be discussed in later posts.
Node and Cluster Layout
GPU-Friendly Design
Unlike typical spiking neural nets, EvoNet is built for GPUs by:
Iteratively updating the entire network in parallel via sparse matrix multiplications.
Exploiting sparsity to skip zero connections.
Distributing clusters across GPUs, with only light communication between them. Similar to model parallelism.
This should make it computationally plausible even on existing hardware.
Next Steps: Testing in DOOM
I plan to continue by testing EvoNet in DOOM or another simple RL environment. These games are perfect early testbeds: they’re complex enough to challenge learning, but controlled enough to clearly observe what the network is doing.
The main goal isn’t just to perform well, but to see whether EvoNet’s iterative, state-driven architecture can genuinely adapt over time, building stable internal representations and handling new situations without brute-force parameter scaling.
Call for Collaboration & Support
I am an independent researcher working on this without institutional backing. Funding is currently the most significant hurdle — even modest resources would let me run experiments that are not possible on personal hardware.
I mainly wanted to publish this so the idea exists publicly; even if I cannot fully pursue it, it might inspire others to build on it. (I've also already written up detailed sections on entropy-guided learning, self-supervision, alignment, and expansion, which I plan to post if there is interest.)
If you are a researcher, engineer, hobbyist, or simply curious about this topic, I would love to hear your thoughts, critiques, or any leads on potential collaboration. Even small contributions can help bring this very different kind of architecture closer to reality.
I have read some prior posts on spiking neural nets and graph-based models here, but have not seen detailed explorations of persistent, iterative flows like this. If you know of discussions or papers that explore similar architectures, I would love pointers.
Can such persistent, iterative flows in a loosely structured graph really carry and refine useful information, or is this more likely to collapse under noise or uniformity?