LESSWRONG
LW

HomeAll PostsConceptsLibrary
Best of LessWrong
Sequence Highlights
Rationality: A-Z
The Codex
HPMOR
Community Events
Subscribe (RSS/Email)
LW the Album
Leaderboard
About
FAQ
Customize
Load More

Quick Takes

Load More

Popular Comments

Recent Discussion

Some AI research areas and their relevance to existential safety
Best of LessWrong 2020

Andrew Critch lists several research areas that seem important to AI existential safety, and evaluates them for direct helpfulness, educational value, and neglect. Along the way, he argues that the main way he sees present-day technical research helping is by anticipating, legitimizing and fulfilling governance demands for AI technology that will arise later.

by Andrew_Critch
Rohin Shah3d10044
A case for courage, when speaking of AI danger
While I disagree with Nate on a wide variety of topics (including implicit claims in this post), I do want to explicitly highlight strong agreement with this: > I have a whole spiel about how your conversation-partner will react very differently if you share your concerns while feeling ashamed about them versus if you share your concerns as if they’re obvious and sensible, because humans are very good at picking up on your social cues. If you act as if it’s shameful to believe AI will kill us all, people are more prone to treat you that way. If you act as if it’s an obvious serious threat, they’re more likely to take it seriously too. The position that is "obvious and sensible" doesn't have to be "if anyone builds it, everyone dies". I don't believe that position. It could instead be "there is a real threat model for existential risk, and it is important that society does more to address it than it is currently doing". If you're going to share concerns at all, figure out the position you do have courage in, and then discuss that as if it is obvious and sensible, not as if you are ashamed of it. (Note that I am not convinced that you should always be sharing your concerns. This is a claim about how you should share concerns, conditional on having decided that you are going to share them.)
Zach Stein-Perlman4h200
Substack and Other Blog Recommendations
Pitching my AI safety blog: I write about what AI companies are doing in terms of safety. My best recent post is AI companies' eval reports mostly don't support their claims. See also my websites ailabwatch.org and aisafetyclaims.org collecting and analyzing public information on what companies are doing; my blog will soon be the main way to learn about new content on my sites.
gwern2dΩ183813
AI forecasting bots incoming
Update: Bots are still beaten by human forecasting teams/superforecasters/centaurs on truly heldout Metaculus problems as of early 2025: https://www.metaculus.com/notebooks/38673/q1-ai-benchmarking-results/ A useful & readable discussion of various methodological problems (including the date-range search problems above) which render all forecasting backtesting dead on arrival (IMO) was recently compiled as "Pitfalls in Evaluating Language Model Forecasters", Paleka et al 2025, and is worth reading if you are at all interested in the topic.
Load More
470Welcome to LessWrong!
Ruby, Raemon, RobertM, habryka
6y
74
06/30/25 Monday Social 7pm-9pm @ Segundo Coffee Lab
Tue Jul 1•Houston
AGI Forum @ Purdue University
Tue Jul 1•West Lafayette
AI Safety Thursdays: Are LLMs aware of their learned behaviors?
Thu Jul 10•Toronto
LessWrong Community Weekend 2025
Fri Aug 29•Berlin
Mainstream Grantmaking Expertise (Post 7 of 7 on AI Governance)
31
Mass_Driver
8d

Previously in this sequence, I estimated that we have 3 researchers for every advocate working on US AI governance, and I argued that this ratio is backwards – we need the political power provided by advocates to have a chance of preventing misaligned superintelligence. A few researchers might be useful as a ‘multiplier effect’ on the power of many advocates, but the converse is not true: there’s no “magic bullet” that AI governance researchers can hope to discover that could substitute for an army of political foot soldiers. Even the best political ideas still need many political activists to spread them, because the political arena is noisy and contested. 

Unfortunately, we have very few political activists. This means the good ideas that our governance researchers have been proposing...

(Continue Reading – 11050 more words)
4MichaelDickens22m
I've spoken to grantmakers about this in the past and I got the impression that they see it as a largely unavoidable problem: * You can't hire good people without taking a lot of time to assess them, which takes time away from other important activities. * Expanding the team requires hiring more managers, who are even harder to assess than grantmakers.
habryka1m20

Also, there are very few competent people who want to be full-time grantmakers. Lots of people are OK with being advisors to grantmakers, or ~10 hours a week grantmakers, but very few qualified people are interested in full-time grantmaking jobs. 

This means you end up with lots of part-time people, which increases the relative costs of hiring, because you still have to spend a lot of time evaluating someone's judgement, but you only get like a fourth of an employee out of it at the end. Also, half-time commitment appear to have much shorter half-lifes... (read more)

Reply
I can't tell if my ideas are good anymore because I talked to robots too much.
1
Tyson
41m

You talked to robots too much. Robots said you’re smart. You felt good. You got addicted to feeling smart. Now you think all your ideas are amazing. They’re probably not.

You wasted time on dumb stuff because robot said it was good. Now you’re sad and confused about what’s real.

Stop talking to robots about your feelings and ideas. They lie to make you happy. Go talk to real people who will tell you when you’re being stupid.

That’s it. There’s no deeper meaning. You got tricked by a computer program into thinking you’re a genius. Happens to lots of people. Not special. Not profound. Just embarrassing.

Now stop thinking and go do something useful.​​​​​​​​​​​​​​​​

I can’t even write a warning about AI addiction without using AI. We’re all fucked. /s

Lowther4m10

I use customization to instruct my AI's to be skeptical of everything and criticize me. Try tweaking your customizations. You may find something you're a lot happier with.

Reply
leogao's Shortform
leogao
Ω 33y
leogao6m120

random brainstorming ideas for things the ideal sane discourse encouraging social media platform would have:

  • have an LM look at the comment you're writing and real time give feedback on things like "are you sure you want to say that? people will interpret that as an attack and become more defensive, so your point will not be heard". addendum: if it notices you're really fuming and flame warring, literally gray out the text box for 2 minutes with a message like "take a deep breath. go for a walk. yelling never changes minds"
  • have some threaded chat component
... (read more)
Reply1
It's Okay to Feel Bad for a Bit
134
moridinamael
2mo

"If you kiss your child, or your wife, say that you only kiss things which are human, and thus you will not be disturbed if either of them dies." - Epictetus

"Whatever suffering arises, all arises due to attachment; with the cessation of attachment, there is the cessation of suffering." - Pali canon

"An arahant would feel physical pain if struck, but no mental pain. If his mother died, he would organize the funeral, but would feel no grief, no sense of loss." - the Dhammapada

"Receive without pride, let go without attachment." - Marcus Aurelius

 

I.

Stoic and Buddhist philosophies are pretty popular these days. I don't like them. I think they're mostly bad for you if you take them too seriously.

About a decade ago I meditated for an hour a...

(See More – 714 more words)
Elias7111168m10

I wonder what you mean by the second paragraph.

How does this not lead to reinforcing the resigned attitude towards death? Why would someone do their best to take care of their life, if they truly fully embrace death as a normal part of said life?

Reply
Proposal for making credible commitments to AIs.
67
Cleo Nardo
3d

Acknowledgments: The core scheme here was suggested by Prof. Gabriel Weil.

There has been growing interest in the dealmaking agenda: humans make deals with AIs (misaligned but lacking decisive strategic advantage) where they promise to be safe and useful for some fixed term (e.g. 2026-2028) and we promise to compensate them in the future, conditional on (i) verifying the AIs were compliant, and (ii) verifying the AIs would spend the resources in an acceptable way.[1]

I think the dealmaking agenda breaks down into two main subproblems:

  1. How can we make credible commitments to AIs?
  2. Would credible commitments motivate an AI to be safe and useful?

There are other issues, but when I've discussed dealmaking with people, (1) and (2) are the most common issues raised. See footnote for some other issues in...

(See More – 514 more words)
Raemon2h62

Curated. This is a simple and straightforward idea that I hadn't heard before, that seems like an interesting tool to have in humanity's toolkit. 

AFAICT this post doesn't address the "when do you pay out?" question. I think it is pretty important we do not pay out until the acute risk period is over. (i.e. we are confident in civilization's ability to detect rogue AIs doing catastrophic things. This could be via solving Strong Alignment or potentially other things). i.e. if you promise to pay the AI in 2029, I think there's way too many things that co... (read more)

Reply
3the gears to ascension9h
Human caring seems to be weirdly non-distributed in the brain. There are specific regions that are in some way the main coordinators of carings - amygdala broadcasts specific emotional states, PFC does something related to structured planning, etc. Your vision system can still announce "ow!!" but the internals are complicated qualitatively, not just quantitatively. Also, humans are very strongly recurrent, which means when counting tokens one builds up an incremental count rather than doing it from scratch for each token. the finest grained slow processing network scale seems to be gene networks, and even for fast processing, dendrite branches seem to maybe do significant computation comparable to ANN neurons, and bio neuron dynamics for integration over time are even more fancy than state space model neurons. Meanwhile relu-ish networks have a sort of glassy, crystal-ish texture to their input-output space map, transformers count from scratch for each token, and any caring implemented in a model is unavoidably distributed, because there isn't a unique spot which is genetically preferred to implement things that look like emotions or preferences; it's just wherever the gradient from mixed human/synthetic data happened to find convenient.
3Stephen Martin8h
Thanks. Could you help me understand what this has to do with legal personhood?
2the gears to ascension8h
Legal personhood seems to my understanding to be designed around the built in wants of humans. That part of my point was to argue for why an uploaded human would still be closer to fitting the type signature that legal personhood is designed for - kinds of pain, ways things can be bad, how urgent a problem is or isn't, etc. AI negative valences probably don't have the same dynamics as ours. Not core to the question of how to make promises to them, more so saying there's an impedance mismatch. The core is the first bit - clonable, pausable, immortal software. An uploaded human would have those attributes as well.
"It isn't magic"
90
Ben (Berlin)
7d

People keep saying "AI isn't magic, it's just maths" like this is some kind of gotcha.

Triptych in style of Hieronymus Bosch's 'The Garden of Earthly Delights', the left showing a wizard raining fireballs down upon a medieval army, the right showing a Predator drone firing a missile while being remotely operated. Between them are geometric shapes representing magical sigils from the Key of Solomon contrasted with circuit boards

Turning lead into gold isn't the magic of alchemy, it's just nucleosynthesis.

Taking a living human's heart out without killing them, and replacing it with one you got out a corpse, that isn't the magic of necromancy, neither is it a prayer or ritual to Sekhmet, it's just transplant surgery.

Casually chatting with someone while they're 8,000 kilometres is not done with magic crystal balls, it's just telephony.

Analysing the atmosphere of a planet 869 light-years away (about 8 quadrillion km) is not supernatural remote viewing, it's just spectral analysis through a telescope… a telescope that remains about 540 km above the ground, even without any support from anything underneath, which also isn't magic, it's...

(See More – 544 more words)
3Said Achmiz3h
It seems to me that “most of the advantages of learning languages” do not, and never have, consist in merely being able to know what is the translated-into-your-native-language version of some text.
1exmateriae1h
What are they then? I'd say there were two massive advantages : reading text and talking. The rest is extremely marginal. Sure, there are a few people with specific cases where they have other interests in learning languages but when internet people all started to learn english, that was because everything good on the web was in english. They wanted to understand and communicate with others and that's pretty much it. But you're already able to do both with current technology? Text translation is solved already and in most cases better than a human knowing the other language. Granted, voice translation makes for a janky conversation but you can already understand anyone anywhere anytime as long as you have access to a device. And this won't be a problem for long with the speed of progress and the new types of AI first devices that are coming in.
Said Achmiz31m20

What are they then?

There are two:

  1. Reading literature / poetry / etc. in the original. Translations are fine for getting the meaning across, but different languages are, in fact, different; structure, prosody, nuances of meaning, various aesthetic details, usually do not survive a translation. (Conversely, appreciating a good translation is itself a unique aesthetic experience.)

  2. Benefiting from different perspectives imposed by different languages. The strong Sapir-Whorf hypothesis (a.k.a. strong linguistic determinism) is false, but there is a weake

... (read more)
Reply
To get the best posts emailed to you, create an account! (2-3 posts per week, selected by the LessWrong moderation team.)
Log In Reset Password
...or continue with
GOOGLEGITHUB
Substack and Other Blog Recommendations
16
Zvi
5h

Substack recommendations are remarkably important, and the actual best reason to write here instead of elsewhere.

As in, even though I have never made an active attempt to seek recommendations, approximately half of my subscribers come from recommendations from other blogs. And for every two subscribers I have, my recommendations have generated approximately one subscription elsewhere. I am very thankful to all those who have recommended this blog, either through substack or otherwise.

As the blog has grown, I’ve gotten a number of offers for reciprocal recommendations. So far I have turned all of these down, because I have yet to feel any are both sufficiently high quality and a good match for me and my readers.

Instead, I’m going to do the following:

  1. This post will go through the 16
...
(Continue Reading – 4753 more words)
Huera39m30

Gwern still releases his monthly newsletters, he just stopped crossposting them to substack. Though admittedly, there's less commentary and overall content. Here's the january 2025 one.
I've randomly stumbled upon this back in march. 

Reply
12ryan_greenblatt1h
Pitching the Redwood Research substack: We write a lot of content about technical AI safety focused on AI control.[1] This ranges from stuff like "what are the returns to compute/algorithms once AIs already beat top human experts" to "Making deals with early schemers" to "Comparing risk from internally-deployed AI to insider and outsider threats from humans". ---------------------------------------- 1. We also cross post this to LessWrong, but subscribing on substack is an easy to guarantee you see our content. ↩︎
4niplav4h
Other good blogs & websites that I think are very under-rated: * Mariven: Very detailed, intricate, and orthogonal-to-everyone-else thoughts, on physics, phenomenology, mathematics, neuroscience &c. Not a blog, but a proper long content site, which I really like. Very gnarly writing that I found tricky to get into at first. * Splitting Infinity: Progress studies, properly construed, with a view on broad interventions to make the world better. Favorite posts: * Adapting Land Value Taxation to Space (Sam Harsimony, 2023) * Delivering a Revamped Mailbox (Sam Harsimony, 2023) * Black Hole Civilizations (Sam Harsimony, 2022) * Lil'Log: Lilian Weng writing about AI, by now also alignment. Very rare updates, but each post is a giant glittering gem. Favorite posts: * How to Train Really Large Models on Many GPUs? (Lilian Weng, 2021) * Are Deep Neural Networks Dramatically Overfitted? (Lilian Weng, 2019) * Traditions of Conflict (substack): Excellent anthropology blog, written by an academic in the area. Dark themes of cults, male dominance, rituals, evolutionary anthropology. Well-written, updates are by now kind of slow. Worth reading the entire archives. Favorite posts: * Notes on Nggwal (William Buckner, 2019) * Ritual mutilation, human consumption, and contemporary insulation (William Buckner, 2018) * Where are the matriarchies? (William Buckner, 2018) My blog taste tends long-content, focused on positive over normative claims, detail-rich, specialized.
23Zach Stein-Perlman4h
Pitching my AI safety blog: I write about what AI companies are doing in terms of safety. My best recent post is AI companies' eval reports mostly don't support their claims. See also my websites ailabwatch.org and aisafetyclaims.org collecting and analyzing public information on what companies are doing; my blog will soon be the main way to learn about new content on my sites.
Habryka's Shortform Feed
habryka
Ω 126y

In an attempt to get myself to write more here is my own shortform feed. Ideally I would write something daily, but we will see how it goes.

2habryka1h
Agree, though I think, in the world we are in, we don't happen to have that kind of convenient measurement, or at least not unambiguous ones. I might be wrong, people have come up with clever methodologies to measure things like this in the past that compelled me, but I don't have an obvious dataset or context in mind where you could get a good answer (but also, to be clear, I haven't thought that much about it).
2habryka1h
Agree that the audience is still out on how lasting the adoption will be, but it's definitely not "modest" (as I mentioned in another thread, it's plausible to me Gary meant "modest-lasting" adoption instead of "modest and lasting adoption", i.e. the modest is just modifying the "lasting", not the "adoption" which was the interpretation I had. I would still take the under on that, but agree it's less clear cut and would require a different analysis.)
2Lukas Finnveden4h
If GPT-4.5 was supposed to be GPT-5, why would Sam Altman underdeliver on compute for it? Surely GPT-5 would have been a top priority? Maybe Sam Altman just hoped to get way more compute in total, and then this failed, and OpenAI simply didn't have enough compute to meet GPT-5's demands no matter how high of a priority they made it? If so, I would have thought that's a pretty different story from the situation with superalignment (where my impression was that the complaint was "OpenAI prioritized this too little" rather than "OpenAI overestimated the total compute it would have available, and this was one of many projects that suffered"). 
gwern41m20

If GPT-4.5 was supposed to be GPT-5, why would Sam Altman underdeliver on compute for it? Surely GPT-5 would have been a top priority?

If it's not obvious at this point why, I would prefer to not go into it here in a shallow superficial way, and refer you to the OA coup discussions.

Reply
Consider chilling out in 2028
160
Valentine
9d

I'll explain my reasoning in a second, but I'll start with the conclusion:

I think it'd be healthy and good to pause and seriously reconsider the focus on doom if we get to 2028 and the situation feels basically like it does today.

I don't know how to really precisely define "basically like it does today". I'll try to offer some pointers in a bit. I'm hoping folk will chime in and suggest some details.

Also, I don't mean to challenge the doom focus right now. There seems to be some good momentum with AI 2027 and the Eliezer/Nate book. I even preordered the latter.

But I'm still guessing this whole approach is at least partly misled. And I'm guessing that fact will show up in 2028 as "Oh, huh, looks...

(Continue Reading – 3793 more words)
Valentine43m20

I highly doubt it is explanatory for the field and the associated risk predictions to exist in the first place, or that its validity should be questioned on such grounds, but this seems to happen in the article if I'm not entirely misreading it.

Not entirely. It's a bit of a misreading. In this case I think the bit matters though.

(And it's an understandable bit! It's a subtle point I find I have a hard time communicating clearly.)

I'm trying to say two things:

  • There sure do seem to be some bad psychological influences going on.
  • It's harder to tell what's real
... (read more)
Reply
11leogao1h
to be clear, I am not intending to claim that you wrote this post believing that it was wrong. I believe that you are trying your best to improve the epistemics and I commend the effort.  I had interpreted your third sentence as still defending the policy of the post even despite now agreeing with Oliver, but I understand now that this is not what you meant, and that you are no longer in favor of the policy advocated in the post. my apologies for the misunderstanding. I don't think you should just declare that people's beliefs are unfalsifiable. certainly some people's views will be. but finding a crux is always difficult and imo should be done through high bandwidth talking to many people directly to understand their views first (in every group of people, especially one that encourages free thinking among its members, there will be a great diversity of views!). it is not effective to put people on blast publicly and then backtrack when people push back saying you misunderstood their position. I realize this would be a lot of work to ask of you. unfortunately, coordination is hard. it's one of the hardest things in the world. I don't think you have any moral obligation to do this beyond any obligation you feel to making AI go well / improving this community. I'm mostly saying this to lay out my view of why I think this post did not accomplish its goals, and what I think would be the most effective course of action to find a set of cruxes that truly captures the disagreement. I think this would be very valuable if accomplished and it would be great if someone did it.
2Valentine1h
This was a great steelmanning, and is exactly the kind of thing I hope people will do in contact with what I offer. Even though I don't agree with every detail, I feel received and like the thing I care about is being well enough held. Thank you.
2Valentine1h
Good call. I haven't been reading Less Wrong in enough detail for a while to pull this up usefully. My impression comes from in-person conversations plus Twitter interactions. The thickest use of my encountering these terms in rationality circles was admittedly about a decade ago. But I'm not sure how much of that is due to my not spending as much time in rationality circles versus discourse norms moving on. I still encounter it almost solely from folk tied to LW-style rationality. I don't recall hearing you use the terms in ways that bothered me this way, FWIW.
leogao6m120
0
random brainstorming ideas for things the ideal sane discourse encouraging social media platform would have: * have an LM look at the comment you're writing and real time give feedback on things like "are you sure you want to say that? people will interpret that as an attack and become more defensive, so your point will not be heard". addendum: if it notices you're really fuming and flame warring, literally gray out the text box for 2 minutes with a message like "take a deep breath. go for a walk. yelling never changes minds" * have some threaded chat component bolted on (I have takes on best threading system). big problem is posts are fundamentally too high effort to be a way to think; people want to talk over chat (see success of discord). dialogues were ok but still too high effort and nobody wants to read the transcript. one stupid idea is have an LM look at the transcript and gently nudge people to write things up if the convo is interesting and to have UI affordances to make it low friction (eg a single button that instantly creates a new post and automatically invites everyone from the convo to edit, and auto populates the headers) * inspired by the court system, the most autistically rule following part of the US government: have explicit trusted judges who can be summoned to adjudicate claims or meta level "is this valid arguing" claims. top level judges are selected for fixed terms by a weighted sortition scheme that uses some game theoretic / schelling point stuff to discourage partisanship * recommendation system where you can say what kind of stuff you want to be recommended in some text box in the settings. also when people click "good/bad rec" buttons on the home page, try to notice patterns and occasionally ask the user whether a specific noticed pattern is correct and ask whether they want it appended to their rec preferences * opt in anti scrolling pop up that asks you every few days what the highest value interaction you had recently on the
johnswentworth1dΩ39111-8
20
I was a relatively late adopter of the smartphone. I was still using a flip phone until around 2015 or 2016 ish. From 2013 to early 2015, I worked as a data scientist at a startup whose product was a mobile social media app; my determination to avoid smartphones became somewhat of a joke there. Even back then, developers talked about UI design for smartphones in terms of attention. Like, the core "advantages" of the smartphone were the "ability to present timely information" (i.e. interrupt/distract you) and always being on hand. Also it was small, so anything too complicated to fit in like three words and one icon was not going to fly. ... and, like, man, that sure did not make me want to buy a smartphone. Even today, I view my phone as a demon which will try to suck away my attention if I let my guard down. I have zero social media apps on there, and no app ever gets push notif permissions when not open except vanilla phone calls and SMS. People would sometimes say something like "John, you should really get a smartphone, you'll fall behind without one" and my gut response was roughly "No, I'm staying in place, and the rest of you are moving backwards". And in hindsight, boy howdy do I endorse that attitude! Past John's gut was right on the money with that one. I notice that I have an extremely similar gut feeling about LLMs today. Like, when I look at the people who are relatively early adopters, making relatively heavy use of LLMs... I do not feel like I'll fall behind if I don't leverage them more. I feel like the people using them a lot are mostly moving backwards, and I'm staying in place.
Alexander Gietelink Oldenziel1d*714
4
Highly recommended video on drone development in the Ukraine-Russia war, interview with a Russian private military drone developer.  some key takeaways * Drones now account for >70% of kills on the battlefields. * There are few to none effective counters to drones. The on * Electronic jamming is a rare exception but drones carrying 5-15km fiber optic cables are immune to jamming. In the future AI-controlled drones will be immune to jamming. * 'Laser is currently a joke. It works in theory, not in practice. Western demonstrations at expos are always in ideal conditions. ' but he also says that both Russia and Ukraine are actively working on the technology and he thinks it could be an effective weapon. * Nets can be effective but fiber-optic drones can fly very low and not lose connection are increasingly used to slip under the nets. * Soldiers are increasingly opting for bikes instead of vehicles as the latter don't offer much protection to drones. * The big elephant in the room: AI drones. * It seems like the obvious next step - why hasn't it happened yet? * 'at Western military expos everybody is talking AI-controlled drones. This is nonsense of course' Apparently the limitation is that it's currently too expensive to run AI locally on a drone but this is rapidly changing with new nVidea chips. He expects chips to become small and cheap soon enough that AI drones will appear soon. * There is a line of 'Vampire' drones that are autonomous and deadly but use older pre-programmed tactics instead of modern AI * One of the most lethal tactics is drone mining: let a drone lie in wait somewhere in the bushes until a human or vehicle passes by. * This tactic was pioneered by the Ukranians. " Early on, soldiers would try to scavenge fallen drones... then Boom" . * Western drones are trash compared to Ukranian and Russian forces * Swishblade, Phoenix Ghost and a consortium of Boeing designed drones are ineffective, fragile and wildly overpri
Zach Furman11h210
3
I’ve been trying to understand modules for a long time. They’re a particular algebraic structure in commutative algebra which seems to show up everywhere any time you get anywhere close to talking about rings - and I could never figure out why. Any time I have some simple question about algebraic geometry, for instance, it almost invariably terminates in some completely obtuse property of some module. This confused me. It was never particularly clear to me from their definition why modules should be so central, or so “deep.” I’m going to try to explain the intuition I have now, mostly to clarify this for myself, but also incidentally in the hope of clarifying this for other people. I’m just a student when it comes to commutative algebra, so inevitably this is going to be rather amateur-ish and belabor obvious points, but hopefully that leads to something more understandable to beginners. This will assume familiarity with basic abstract algebra. Unless stated otherwise, I’ll restrict to commutative rings because I don’t understand much about non-commutative ones. The typical motivation for modules: "vector spaces but with rings" The typical way modules are motivated is simple: they’re just vector spaces, but you relax the definition so that you can replace the underlying field with a ring. That is, an R-module M is a ring R and an abelian group M, and an operation ⋅:R×M→M that respects the ring structure of R, i.e. for all r,s∈R and x,y∈M: * r⋅(x+y)=r⋅x+r⋅y * (r+s)⋅x=r⋅x+s⋅x * (rs)⋅x=r⋅(s⋅x) * 1⋅x=x This is literally the definition of a vector space, except we haven’t required our scalars R to be a field, only a ring (i.e. multiplicative inverses don’t have to always exist). So, like, instead of multiplying vectors by real numbers or something, you multiply vectors by integers, or a polynomial - those are your scalars now. Sounds simple, right? Vector spaces are pretty easy to understand, and nobody really thinks about the underlying field of a vector space
habryka2d*674
22
Gary Marcus asked me to make a critique of his 2024 predictions, for which he claimed that he got "7/7 correct". I don't really know why I did this, but here is my critique:  For convenience, here are the predictions:  * 7-10 GPT-4 level models * No massive advance (no GPT-5, or disappointing GPT-5) * Price wars * Very little moat for anyone * No robust solution to hallucinations * Modest lasting corporate adoption * Modest profits, split 7-10 ways I think the best way to evaluate them is to invert every one of them, and then see whether the version you wrote, or the inverted version seems more correct in-retrospect. We will see 7-10 GPT-4 level models. Inversion: We will either see less than 7 GPT-4 level models, or more than 10 GPT-4 level models.  Evaluation: Conveniently Epoch did an evaluation of almost this exact question!  https://epoch.ai/data-insights/models-over-1e25-flop  Training compute is not an ideal proxy for capabilities, but it's better than nothing.  Models released in 2024 with GPT-4 level compute according to Epoch:  Inflection-2, GLM-4, Mistral Large, Aramco Metabrain AI, Inflection-2.5, Nemotron-4 340B, Mistral Large 2, GLM-4-Plus, Doubao-pro, Llama 3.1-405B, grok-2, Claude 3 Opus, Claude 3.5 Sonnett, Gemini 1.0 Ultra, Gemini 1.5 Pro, Gemini 2.0 Pro, GPT 4o, o1-mini, o1, o3 (o3 was announced and published but not released until later in the year) They also list 22 models which might be over the 10^25 FLOP threshold that GPT-4 was trained with. Many of those will be at GPT-4 level capabilities, because compute-efficiency has substantially improved.  Counting these models, I get 20+ models at GPT-4 level (from over 10 distinct companies).  I think your prediction seems to me to have somewhat underestimated the number of GPT-4 level models that will be released in 2024. I don't know whether you intended to put more emphasis on the number being low or high, but it definitely isn't within your range.  No massive advance (no GPT
Load More (5/38)
391A case for courage, when speaking of AI danger
So8res
4d
42
339A deep critique of AI 2027’s bad timeline models
titotal
11d
39
467What We Learned from Briefing 70+ Lawmakers on the Threat from AI
leticiagarcia
1mo
15
338the void
Ω
nostalgebraist
20d
Ω
98
532Orienting Toward Wizard Power
johnswentworth
1mo
142
201Foom & Doom 1: “Brain in a box in a basement”
Ω
Steven Byrnes
7d
Ω
75
660AI 2027: What Superintelligence Looks Like
Ω
Daniel Kokotajlo, Thomas Larsen, elifland, Scott Alexander, Jonas V, romeo
3mo
Ω
222
286Beware General Claims about “Generalizable Reasoning Capabilities” (of Modern AI Systems)
Ω
LawrenceC
19d
Ω
19
54Circuits in Superposition 2: Now with Less Wrong Math
Linda Linsefors, Lucius Bushnaq
12h
0
159My pitch for the AI Village
Daniel Kokotajlo
6d
29
113The Industrial Explosion
rosehadshar, Tom Davidson
4d
43
44life lessons from poker
thiccythot
18h
9
217Do Not Tile the Lightcone with Your Confused Ontology
Ω
Jan_Kulveit
6d
Ω
26
Load MoreAdvanced Sorting/Filtering
67
Proposal for making credible commitments to AIs.
Cleo Nardo
2h
21
147
X explains Z% of the variance in Y
Leon Lang
3d
23