I'm unclear on whether the 'dimensionality' (complexity) component to be minimized needs revision from the naive 'number of nonzeros' (or continuous but similar zero-rewarded priors on parameters).
Either:
Does this seem fair?
This appears to be a high-quality book report. Thanks. I didn't see anywhere the 'because' is demonstrated. Is it proved in the citations or do we just have 'plausibly because'?
Physics experiences in optimizing free energy have long inspired ML optimization uses. Did physicists playing with free energy lead to new optimization methods or is it just something people like to talk about?
This kind of reply is ridiculous and insulting.
We have good reason to suspect that biological intelligence, and hence human intelligence roughly follow similar scaling law patterns to what we observe in machine learning systems
No, we don't. Please state the reason(s) explicitly.
Google's production search is expensive to change, but I'm sure you're right that it is missing some obvious improvements in 'understanding' a la ChatGPT.
One valid excuse for low quality results is that Google's method is actively gamed (for obvious $ reasons) by people who probably have insider info.
IMO a fair comparison would require ChatGPT to do a better job presenting a list of URLs.
how is a discretized weight/activation set amenable to the usual gradient descent optimizers?
You have the profits from the AI tech (+ compute supporting it) vendors and you have the improvements to everyone's work from the AI. Presumably the improvements are more than the take by the AI sellers (esp. if open source tools are used). So it's not appropriate to say that a small "sells AI" industry equates to a small impact on GDP.
But yes, obviously GDP growth climbing to 20% annually and staying there even for 5 years is ridiculous unless you're a takeoff-believer.
You don't have to compute the rotation every time for the weight matrix. You can compute it once. It's true that you have to actually rotate the input activations for every input but that's really trivial.
Having a large pool of specific info on effective recall is a sign of mental health and quite useful. I've noticed various successful and charismatic commentators appearing to have talent in this area. It's possible that as well as being a sign of health it buffers brain abilities generally, that modern recall-augmenting tools will atrophy the native facility. It seems you can IQ test pretty high as long as you're capable of remembering what words mean but otherwise aren't guaranteed to have exceptional long-term memory capacity.