LESSWRONG
LW

6
Nicholas Kross
2072Ω2315220
Message
Dialogue
Subscribe

Theoretical AI alignment (and relevant upskilling) in my free time.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
5NicholasKross's Shortform
3y
19
Before LLM Psychosis, There Was Yes-Man Psychosis
Nicholas Kross22d219

I think lots of humans are also just starved for compliments in general, outside of contexts like "You are a waiter and the social script tells the table to thank you" or "You are someone's spouse". The classic example in my mind is "male vs female socialization" and its consequences, e.g. as discussed in this screenshotted tumblr post or this other one. Men often, in some sense, get "too few" wanted compliments, and women get inundated with too many unwanted ones. "It's like one person dying of dehydration watching another one drown". Then, of course, classic nightmarish social/cultural/internet incentives layer atop this dynamic and make it worse.

Widespread chronic human under-complimented-ness seems relatively easy to explain, from the supply-side. How often does the average person hand out unsolicited compliments, outside of well-known contexts like the restaurant example above? I'd hazard a guess of "too little". People could easily be "well-put-together" and still suffering from this, just as "feeling full" and "having proper nutrition" aren't the same.

Reply
Ryan Kidd's Shortform
Nicholas Kross23d20

I'm curious to see if I'm in this data, so I can help make it more accurate by providing info.

Reply
ZY's Shortform
Nicholas Kross3mo20

https://www.lesswrong.com/posts/tpZciMYCXN49FYWnS/nicholaskross-s-shortform?commentId=f4PxFp8LkKKxdCXxh

Reply1
Eliezer and I wrote a book: If Anyone Builds It, Everyone Dies
Nicholas Kross3mo20

From the MIRI announcement:

Our big ask for you is: If you have any way to help this book do shockingly, absurdly well— in ways that prompt a serious and sober response from the world — then now is the time.

sober response from the world

sober response

Uh... this is debatably a lot to ask of the world right now.

Reply
Rationalist Movie Reviews
Nicholas Kross7mo20

I said "one of the best movies about", not "one of the best movies showing you how to".

Reply
NicholasKross's Shortform
Nicholas Kross9mo31

The punchline is "alignment could productively use more funding". Many of us already know that, but I felt like putting a mildly-opinionated spin on what kind of things, at the margin, may help top researchers. (Also I spent several minutes editing/hedging the joke)

Reply
NicholasKross's Shortform
Nicholas Kross9mo31

Virgin 2030s [sic] MIRI fellow:
- is cared for so they can focus on research
- has staff to do their laundry
- soyboys who don't know *real* struggle
- 3 LDT-level alignment breakthroughs per week

CHAD 2010s Yudkowsky:
- founded a whole movement to support himself
- "IN A CAVE, WITH A BOX OF SCRAPS"
- walked uphill both ways to Lightcone offices.
- alpha who knows *real* struggle
- 1 LDT-level alignment breakthrough per decade

Reply1
Load More
16Rationalist Movie Reviews
7mo
2
9Is principled mass-outreach possible, for AGI X-risk?
2y
5
16How to Get Rationalist Feedback
2y
0
-13Musk, Starlink, and Crimea
2y
0
28Incentives affecting alignment-researcher encouragement
Q
2y
Q
3
11Build knowledge base first, or backchain?
Q
2y
Q
5
14Rationality, Pedagogy, and "Vibes": Quick Thoughts
2y
1
16How to Search Multiple Websites Quickly
2y
1
41Why I'm Not (Yet) A Full-Time Technical Alignment Researcher
2y
21
263[SEE NEW EDITS] No, *You* Need to Write Clearer
2y
65
Load More