Further signal boost.
I use those playlists a lot. Thank you.
Perhaps your team could have helped write the safety part?
Or to deliberate whether the weights and code should be made public?
The name of the paper is very meaningful (AGA ≈ AGI, obviously on purpose), so in order to get in condition, I think it is important that your safety team takes part in this kind of paper.
Perhaps your team could have helped write the safety part?
I think it would be a bad use of our time to write the safety sections of all the papers that could be progress towards AGI (there are a lot of them). It seems a lot better to focus on generally improving knowledge of safety, and letting individual projects write their own safety sections.
Obviously if an actually x-risky system is being built it would be important for us to be involved but I think this was not particularly x-risky.
Tbc we would have been happy to chat to them if they reached out; I'm... (read more)
Demis Hassabis is not even mentioned in the paper. Does that mean this is considered a minor paper for DeepMind?
There is no fire alarm for AGIs? Maybe just subscribe to the DeepMind RSS feed…
On a more serious note, I'm curious about the internal review process for this article, what role did the DeepMind AI safety team play in it? In the Acknowledgements, there is no mention of their contribution.
I don't think we played any particular role in the review process (or if we did I'm not aware of it, which tbc is totally possible, I don't know everything that everyone on DeepMind safety does).
What would you want us to bring up if we were to participate in a review process?
Ah, "The tool-assisted attack went from taking 13 minutes to taking 26 minutes per example."
Interesting. Changing the in-distribution (3oom) does not influences much the out-distribution (*2)
I do not understand the "we only doubled the amount of effort necessary to generate an adversarial counterexample.". Aren't we talking about 3oom?
Super work! It must have required a crazy amount of technical and manual work.
The fact that you manage to reduce the number of failures by 3 orders of magnitude is quite impressive. Did you separate train and test set at the beginning of the project?
In conclusion, using a 300M parameter model to supervise a 10 times larger model seems to work well, it gives a lot of hope.
I just discovered this debate thanks to a YouTube recommendation from the Institute for the Future of Life. I find the formulation of the question by Anderw Serazin very well put.
The Yes Team: We would identify with it better, and it would help with creativity. You would want your digital assistant to be able to help you by modeling you, it seems necessary to model the human.
No team: It's easier to deal with IA safety without it, the possibility to create infinite suffering is bad, unpleasant task in the world are better done by non-consc... (read more)
The difference between primary consciousness and higher order consciousness is not clear to me. Is it the same thing as access consciousness and meta cognition?
"and higher order consciousness, which came to only humans with the acquisition of language. A machine with primary consciousness will probably have to come first"
The language models like GPT3 appear to be unstoppable, and are much more versatile than the models of other modalities.
So maybe a machine with higher order consciousness will come first?
Yeah, I know, I just wanted to begin answering with this and to present in one sentence (and without mentioning it, my bad...) the concept of neural correlates of consciousness
Are you using "conscious" and "sentient" as synonyms? → Yes. Maybe sentient is consciousness with only valence but it is basically the same thing. But you're right, I shouldn't have used both worlds at the same time without further explanation.
Do you think some, most, or all humans are conscious? → Yes, I think most humans are conscious. Why do you think this? Because I am conscious, but I'm not sure if I really answer your question here.
"I'm not sure what it means for it to be "desirable" for an entity to be conscious." → Yes this is a good qu... (read more)
I agree that today the concept of consciousness is very poorly defined, but I think that in the future it will be possible to define it in a way that will make sense, or at least that we will be able to correct our current intuitions.
How can one tell if a human is conscious?
In humans, we have clues. For example, it is possible to experiment by varying the duration of a light stimulus. There are stimuli of very short duration that are processed by visual areas V1 and V2, but which do not go up to the parietal cortex. For stimuli of slightly longer duration,... (read more)
Either consciousness is a mechanism that has been recruited by evolution for one of its abilities to efficiently integrate information, or consciousness is a type of epiphenomenon that serves no purpose.
Personally I think that consciousness, whatever it is, serves a purpose, and has an importance for the systems that try to sort out the anecdotal information from the information that deserves more extensive consideration. It is possible that this is the only way to effectively process information, and therefore that in trying to program an agi, one naturally comes across it
Interesting.
Just curious, what's the problem with "nu" in verbal conversation ?
Isn't this paper already a shrieking fire alarm?
Visual Studio Code lets you perform most tasks directly from the keyboard. You can even use a Vim emulator if you like.
But more importantly, "faster edits gives me a faster iteration time" : when developing complex stuff, your writing speed is clearly not the limiting factor. Using proper file structure visualization and navigation tools is way more important.
You can SSH directly with VS Code in just one click with the remote extension.
Why would Vim be important ? I mean, everybody uses VS Code nowadays, and it's much more easy and versatile and no need to read a book to understand it...
Could you just explain a bit "will only be likely to contain arbitrary patterns of sizes up to log(10^120)" please ? Or give some pointers with other usage of such calculation ?
This is very much a heuristic, but good enough in this case.
Suppose we want to know how many times we expect to see a pattern with n cells in a random field of area A. Ignoring edge effects, there are A different offsets at which the pattern could appear. Each of these has a 1/2^n chance of being the pattern. So we expect at least one copy of the pattern if n < log_2(A).
In this case the area is (10^60)^2, so we expect patterns of size up to 398.631. In other words, we expect the ash to contain any pattern you can fit in a 20 by 20 box.
Thanks for posting!
upvoted.
But ln(370/20)/ln(2) = 4.2. This means that the new strain doubled 4 times between September and mid-November, suggesting a typical doubling time of just over two weeks.
This is approximately what is observed at the end of December.
But indeed, I don't understand why the number of infected people suddenly decreases at the end of November. An explanation would be helpful.
Where can we find the source saying that there were about 20 cases of new strains in September?
I read this book two years ago when it was published in French. I found it incredibly exciting to read, and that's what motivated me to discover this site and then move on to a master's degree in machine learning.
This book saved me a lot of time in discovering Bayesianism, and made a much deeper change in my way of thinking than if I had simply read a textbook of Bayesian machine learning.
I am of course happy to have read the sequences, but I think I am lucky to have started with the equation of knowledge which is much shorter to read and which provides the theoretical assurances, motivation, main tools, enthusiasm and pedagogy to engage in the quest for Bayesianism.
Good Post!
Here is another point. The population of a city is constrained by the agricultural area accessible in less than 3-4 days, which is the time corresponding to the storage time of vegetables and fruits. During antiquity, Paris was the biggest town in France and was inhabited by 10,000 inhabitants, which corresponds to the population fed by a circle of arable land within a 3-day radius of oxcarts. If in the future transport becomes more constrained (oil shortage?), we should then expect to see the size of the cities greatly reduced.
If you speak French you can look at Jean-Marc Jancovici's lecture https://www.youtube.com/watch?v=Ci_cz18A2F8
Yeah, but writing a sequence seems more fun than doing a literature review.