LESSWRONG
LW

Kabir Kumar
56852390
Message
Dialogue
Subscribe

Running https://aiplans.org 

Fulltime working on the alignment problem. 

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
2Kabir Kumar's Shortform
10mo
155
The Cats are On To Something
Kabir Kumar13h10

I think even among cynics, it's a pretty rare idea that it's that bad

Reply
Kabir Kumar's Shortform
Kabir Kumar3d*10

Poor autistic (or close enough) AI Safety nerds! Come learn how to make money with law! And it's not capabilities!

https://luma.com/8hv5n7t0 

Learn what it looks like to produce an 'expert report' that's actually useful to a law firm. 

(AI Safety researchers will get free entry)

It's a 2 day law hackthon on how enterprise customers can make better contracts or do countersuits with frontier AI companies. Rn, the standard OpenAI contract basically passes all liability onto customers, even for stuff they don't have any control over and should really be OpenAI's responsibility.

technical AI Safety people can give info on stuff like latent dangerous capabilities being easily elictable from most major models, emergent misalignment, etc and help companies save a lot of money. And by learning how to do that, and making contacts during the event, get into positions where they're paid to do this.

Reply1
Generative AI is not causing YCombinator companies to grow more quickly than usual (yet)
Kabir Kumar9d40

I'd be very interested in if this is due to the US economy as a whole being worse now than in 2009. Could we compare with growth rate of AI companies in countries with better economies?

Reply
Kabir Kumar's Shortform
Kabir Kumar11d10

These are imperfect, I'd like feedback on them please:
https://moonshot-alignment-program.notion.site/Proposed-Research-Guides-255a2fee3c6780f68a59d07440e06d53?pvs=74

Reply
Von Neumann's Fallacy and You
Kabir Kumar13d10

I'd also like some citation for this, please.

Reply
Banning Said Achmiz (and broader thoughts on moderation)
Kabir Kumar17d122

My best guess is that the usual ratio of "time it takes to write a critical comment" to "time it takes to respond to it to a level that will broadly be accepted well" is about 5x. This isn't in itself a problem in an environment with lots of mutual trust and trade, but in an adversarial context it means that it's easily possible to run a DDOS attack on basically any author whose contributions you do not like by just asking lots of questions, insinuating holes or potential missing considerations, and demanding a response, approximately independently of the quality of their writing. 

For related musings see the Scott Alexander classic Beware Isolated Demands For Rigor. 

Not strictly related to this post, but I'm glad you know this and it makes me more confident in the future health of Lesswrong as a discussion place.

Reply
Kabir Kumar's Shortform
Kabir Kumar20d31

I think this is easier to anonymize, with the exception of very specific things that people become famous for.

Reply
Kabir Kumar's Shortform
Kabir Kumar20d102

A solution I've come around to for this is retroactive funding. As in, if someone did something essentially without funding, that resulted in outcomes, which if you knew were guaranteed, you would have funded/donated to the project, then donate to the person to encourage them to do it more. 

Reply
Kabir Kumar's Shortform
Kabir Kumar20d30

my mum said to my little sister to take a break from her practice test for the eleven plus and come eat dinner, in the kitchen, with the rest of the family, my gran me and her (dad is upsairs, he's rarely here for family dinner). my little sister, in a trembling voice said 'but then dad will say '

mum sharply says to leave it and come eat dinner. she leaves the living room where my little sister is, goes to the kitchen. my little sister tries to shut off the lights in the living room, when it stutters, beats her little hands onto it in frustration. 

mum hears her in the kitchen, goes to the living room and sharply scolds her 'what will happen if it breaks?? how much will it take to fix it!?'

in the kitchen, in hindi, which my little sister understands just a little bit, my nan lovingly calls my little sister her child, her love, her darling, invites her to sit across from her. 

1 minute later, dad comes down. sees my little sister in the kitchen. starting shouting. "HOW DARE YOU LEAVE YOUR WORK"

"THIS IS YOUR RESPONSIBILITY"

he yells at my mum in hindi that "YOU LOT (referring to my mum and also to me) HAVE MADE THIS A JOKE"

"HOW MANY QUESTIONS HAVE YOU DONE"

"THIS IS NOT EVEN HALF"

"I LEFT HALF AN HOUR AGO"

"STOP!"

"STOP SHAKING LIKE A FISH!"

my sister is sobbing and crying

"STOP CRYING"

"YOU WILL NOT GET FOOD TODAY"

"NO! YOU! WILL! NOT! GET! FOOD! TODAY!"

he goes back upstairs. 

my little sister is sobbing and crying in the living room, minutes later. 

mum is in the kitchen. she knows that if she goes to console my little sister in the living room then my dad will come downstairs and shout again but worse. 

she yells at my nan to be quiet. 

i know that if i go to console my little sister, not only will my dad come downstairs again and shout worse and do worse, he's likely to use this as the excuse to snap and kick me out of the house. and i'm broke, weak and powerless to do much other than write this and beg, hope that someone or that somehow i can change things. i can stop being broke and get a home, a house, where my little sister can be free.

Reply9
Kabir Kumar's Shortform
Kabir Kumar20d10

just joined the call with one of the moonshot teams and i was actually basically an interruption, lol. felt so good to be completely unneeded there

Reply
Load More
19AI Safety Law-a-thon: Turning Alignment Risks into Legal Strategy
5h
0
6Truth
13d
0
71Directly Try Solving Alignment for 5 weeks
2mo
2
2Making progress bars for Alignment
8mo
0
20AI & Liability Ideathon
9mo
2
2Kabir Kumar's Shortform
10mo
155