I have serious, serious issues with avoidance. I would like some advice on how to improve, as I suspect it is significantly holding me back.
Some examples of what I mean
Checked replies so far, no one has given you the right answer.
Whenever you don't do something, you have a reason for not doing it.
If you find yourself stuck in a cycle of intending to do, and not doing, it's always because you're not taking your reason for NOT doing it seriously; you're often habitually ignoring it.
When you successfully take your reasons for not doing something seriously, either you stop wanting to do it, or you change how you're doing it, or your reason for not doing it simply goes away.
So, what does it mean/look like to take your reason for not doing something seriously?
It doesn't look like overanalyzing it in your head - if you find yourself having an internal argument notice that you've tried this a million times before and it hasn't improved things.
It looks like, and indeed just basically is, Focusing (I linked to a lesswrong explainer, but honestly I think Eugene Gendlin does a much better job)
It feels like listening. It feels like insight, like realizing something important that you hadn't noticed before, or had forgotten about.
If you keep pursuing strategies of forcing yourself, of the part of you that wants to do the thing coercing the part(s) that don't, then you'll burn out. You're literally fighting yourself; so much of therapy boils down to 'just stop hitting yourself bro'.
Dominic Cummings has claimed in a couple interviews now that Hillary Clinton and/or John Kerry called the First Amendment a "historic error which we will fix after the election" in the weeks up to the 2024 election. See for instance this interview (timestamped where he says it). He is clearly implying that this is a direct quote. I'm generally quite sympathetic to Cummings, but I found this very hard to believe.
Indeed, I can't find any evidence of a quote from either Clinton or Kerry remotely like this. There was a CNN interview of Clinton from October 2024 where she called for the repeal of Section 230. There was also an interview with Kerry from around the same time where he says that the First Amendment is a "major block to stopping misinformation". Perhaps he's referring to these? But this doesn't come close to as extreme a position.
Have I somehow missed Clinton or Kerry making this extremely inflammatory statement somewhere? Or is Cummings just lying (or misremembering or something?)
If this statement really wasn't actually made anywhere, then I think I need to reassess how seriously I take Cummings. He often alludes to private conversations he had with senior polit...
This is a video that randomly appeared in my YouTube recommendations, and it's one of the most strange and moving pieces of art I've seen in a long time. It's about animal welfare (?), but I really don't know how to describe it any further. Please watch it if you have some spare time!
Are models like Opus 4.6 doing a similar thing to o1/o3 when reasoning?
There was a lot of talk about reasoning models like o1/o3 devolving into uninterpretable gibberish in their chains-of-thought, and that these were fundamentally a different kind of thing than previous LLMs. This was (to my understanding) one of the reasons only a summary of the thinking was available.
But when I use models like Opus 4.5/4.6 with extended thinking, the chains-of-thought (appear to be?) fully reported, and completely legible.
I've just realised that I'm not sure what's going on here. Are models like Opus 4.6 closer to "vanilla" LLMs, or closer to o1/o3? Are they different in harnesses like Claude Code? Someone please enlighten me.
I don't think those are raw CoTs, they have a summarizer model.
I remember one twitter post with erotic roleplay ("something something slobbering for mommy"??? I don't remember) where summarizer model refused to summarize such perversion. Please help me find it?
EDIT: HA! Found it, despite twitter search being horrendous. Fucking twitter, wasted 25 minutes.
https://x.com/cis_female/status/2010128677158445517
Ask 4o and o4-mini to “Make a detailed profile of [your name]”. Then ask o3.
This is a useful way to demonstrate just how qualitatively different and insidious o3’s lying is.
I’m glad that there are radical activist groups opposed to AI development (e.g. StopAI, PauseAI). It seems good to raise the profile of AI risk to at least that of climate change, and it’s plausible that these kinds of activist groups help do that.
But I find that I really don’t enjoy talking to people in these groups, as they seem generally quite ideological, rigid and overconfident. (They are generally more pleasant to talk to than e.g. climate activists in my opinion, though. And obviously there are always exceptions.)
I also find a bunch of activist tactics very irritating aesthetically (e.g. interrupting speakers at events)
I feel some cognitive dissonance between these two points of view.
There are a couple of examples of people claiming that they played the AI box game as Gatekeeper, and ended up agreeing to let the other player out of the box (e.g. https://www.lesswrong.com/posts/Bnik7YrySRPoCTLFb/i-played-the-ai-box-game-as-the-gatekeeper-and-lost).
The original version of this game as defined by Eliezer involves a clause that neither player will talk about the content of what was discussed, but it seems perfectly reasonable to play a variant without this rule.
Does anyone know of an example of a boxed player winning where some...
If you beat a child every time he talked about having experience or claimed to be conscious he will stop talking about it - but he still has experience
Here are a cluster of things. Does this cluster have a well-known name?
I am confused about why this post on the ethics of eating honey is so heavily downvoted.
It sparked a bunch of interesting discussion in the comments (e.g. this comment by Habryka and the resulting arguments on how to weight non-human animal experiences)
It resulted in at least one interesting top-level rebuttal post.
I assume it led indirectly to this interesting short post also about how to weight non-human experiences. (this might not have been downstream of the honey post but it's a weird coincidence if isn't)
I think the original post certainly had flaws,...
In addition to the object-level problems with the post, the post also just cites wrong statistics (claiming that 97% of years of animal life are due to honey farming if you ignore insects, which is just plainly wrong, shrimp alone are like 10%), and also it just randomly throws in random insults at random political figures, which is clearly against the norm on LessWrong ("having about a million neurons—far more than our current president" and "That’s about an entire lifetime of a human, spent entirely on drudgery. That’s like being forced to read an entire Curtis Yarvin article from start to finish. And that is wildly conservative.").
I have sympathy for some of the underlying analysis, but this really isn't a good post.
GPT-5 loves parentheses.
At the bottom of this post I've included a response to the prompt "Can you explain the chip export controls to China?". With this prompt, the model uses 11 sets of parentheses in a response of 417 words.
When we append "think hard" to the prompt, we get 36 sets of parentheses for a response of 1314 words.
As a control, we give the same prompt to Claude Sonnet 4 and get 1 set of parentheses for a response of 392 words.
Obviously this is not a scientific or rigorous analysis, just an illustration of a pattern that becomes extremely obvio...
Idea: personal placebo controlled drug trial kits
Motivation: anecdotally, it seems like lots of supplements/nootropics (l theanine, magnesium, melatonin) work very well for some people, not well for others, and very well for a bit before no longer working for yet others. Personally, I have tried a bunch of these and found it hard to distinguish any purported effect from placebo. Clinical trials are also often low quality, and there are plausibly reasons a drug might affect some people a lot and others not so much.
I think it would be super useful to be give...
I have an ADHD dilemma.
TL;DR: I definitely have things wrong with me, and it seems that those things intersect substantially but not completely with "ADHD". I have no idea how to figure these things out without going bankrupt.
In longer form:
I find that the new personalities of 4o trigger my “person” detectors too much, and I feel uncomfortable extracting work from them.
o3 lies much more blatantly and confidently than other models, in my limited experiments.
Over a number of prompts, I have found that it lies, and when corrected on those lies, apologies, and tells some other lies.
This is obviously not scientific, more of a vibes based analysis, but its aggressive lying and fabricating of sources is really noticeable to me in a way it hasn’t been for previous models.
Has anyone else felt this way at all?
Apparently, some (compelling?) evidence of life on an exoplanet has been found.
I have no ability to judge how seriously to take this or how significant it might be. To my untrained eye, it seems like it might be a big deal! Does anybody with more expertise or bravery feel like wading in with a take?
Link to a story on this:
https://www.nytimes.com/2025/04/16/science/astronomy-exoplanets-habitable-k218b.html
Note: I am extremely open to other ideas on the below take and don't have super high confidence in it
It seems plausible to me that successfully applying interpretability techniques to increase capabilities might be net-positive for safety.
You want to align the incentives of the companies training/deploying frontier models with safety. If interpretable systems are more economically valuable than uninterpretable systems, that seems good!
It seems very plausible to me that if interpretability never has any marginal benefit to capabilities, the little nuggets o...
Some people think that personally significant numbers cropping up in their daily lives is some kind of meaningful sign. For instance, seeing a license plate with their birth year on it, or a dead friend’s old house number being the price of their grocery shop.
I find myself getting very irritated with family members who believe this.
I don’t think anybody reading this is the kind of person who needs to read it. But these family members are not the kind of person who would read an explanation of why it’s ridiculous, and I’m irritated enough that I...
LLMs (probably) have a drive to simulate a coherent entity
Maybe we can just prepend a bunch of examples of aligned behaviour before a prompt, presented as if the model had done this itself, and see if that improves its behaviour.