LESSWRONG
LW

1233
Iknownothing
8916663
Message
Dialogue
Subscribe

Making a research platform for AI Alignment at https://ai-plans.com/
Come critique AI Alignment plans and get feedback on your alignment plan!

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
Even briefer summary of ai-plans.com
Iknownothing2y20

Thank you! Changed it to that!

Reply
AI Law-a-Thon
Iknownothing2y10

Yup, that's definitely something that can be argued by people Against during the Debate Stage!
And they might come to the same conclusion!

Reply
E.T. Jaynes Probability Theory: The logic of Science I
Iknownothing2y20

I'd also read Elementary Analysis before

Reply
E.T. Jaynes Probability Theory: The logic of Science I
Iknownothing2y20

I'm not a grad physics student- I don't have a STEM degree, or the equivalent- I found the book very readable, nonetheless. It's by far my favourite textbook- feels like it was actually written by someone sane, unlike most.

Reply
Critical review of Christiano's disagreements with Yudkowsky
Iknownothing2y33

I'm really glad you wrote this! 
I think you address an important distinction there, but I think there might be a further one to be made- in that how we measure/tell if a model is aligned in the first place. 
There seems to be a growing voice which says that if a model's output seems to be the output we might expect from an aligned AI, then it's aligned. 
I think it's important to distinguish that from the idea that the model is aligned if you actually have a strong idea of what it's values are, how it's gotten them, etc. 

Reply
AI Safety Chatbot
Iknownothing2y10

I'm really excited to see this!! 
I'd like it if this became embed-able so it could be used on ai-plans.com and on other sites!!
Goodness knows, I'd like to be able to get summaries and answers to obscure questions on some alignmentforum posts!

Reply
Why aren't more people in AIS familiar with PDP?
Iknownothing2y10

What do you think someone who knows about PDP knows that someone with a good knowledge of DL doesn't?
And why would it be useful?

Reply
Why Is No One Trying To Align Profit Incentives With Alignment Research?
Iknownothing2y10

I think folks in AI Safety tend to underestimate how powerful and useful liability and an established duty of care would be for this.

Reply
Here's the exit.
Iknownothing2y00

I think calling things a 'game' makes sense to lesswrongers, but just seems unserious to non lesswrongers.

Reply1
How dath ilan coordinates around solving alignment
Iknownothing2y20

I don't think a lack of IQ is the reason we've been failing at making AI sensibly. Rather, it's a lack of good incentive making. 
Making an AI recklessly is current much more profitable than not doing do- which imo, shows a flaw in the efforts which have gone towards making AI safe - as in, not accepting that some people have a very different mindset/beliefs/core values and figuring out a structure/argument that would incentivize people of a broad range of mindsets.

Reply
Load More
24Review of Alignment Plan Critiques- December AI-Plans Critique-a-Thon Results
2y
0
2Proposal for improving state of alignment research
2y
0
6Looking for judges for critiques of Alignment Plans
2y
0
4Specific Arguments against open source LLMs?
Q
2y
Q
2
8AI-Plans.com 10-day Critique-a-Thon
2y
2
4Simple alignment plan that maybe works
2y
8
10Even briefer summary of ai-plans.com
2y
6
2LeCun says making a utility function is intractable
2y
3
9Brief summary of ai-plans.com
2y
4
3An overview of the points system
2y
4
Load More
AI-Plans (website)
2 years ago
(+29)