Kronopath

Posts

Sorted by New

Wiki Contributions

Comments

What would we do if alignment were futile?

Do we have to convince Yann LeCun? Or do we have to convince governments and the public?

(Though I agree that the word "All" is doing a lot of work in that sentence, and that convincing people of this may be hard. But possibly easier than actually solving the alignment problem?)

What would we do if alignment were futile?

A thought: could we already have a case study ready for us?

Governments around the world are talking about regulating tech platforms. Arguably Facebook's News Feed is an AI system and the current narrative is that it's causing mass societal harm due to it optimizing for clicks/likes/time on Facebook/whatever rather than human values.

See also:

All we'd have to do is to convince people that this is actually an AI alignment problem.

Stop button: towards a causal solution

On Wednesday, the lead scientist walks into the lab to discover that the AI has managed to replicate itself several times over, buttons included. The AIs are arranged in pairs, such that each has its robot hand hovering over the button of its partner.

"The AI wasn't supposed to clone itself!" thinks the scientist. "This is bad, I'd better press the stop button on all of these right away!"

At this moment, the robot arms start moving like a swarm of bees, pounding the buttons over and over. If you looked at the network traffic between each computer, you'd see what was happening: the AI kills its partner, then copies itself over to its partner's hard drive, then its partner kills it back, and copies itself back to its original. This happens as fast as the robot arms can move.

Far in the future, the AIs have succeeded in converting 95% of the mass of the earth into pairs of themselves maddeningly pressing each other's buttons and copying themselves as quickly as possible. The only part of the earth that has not been converted into button-pressing AI pairs is a small human oasis, in which the few remaining humans are eternally tortured in the worst way possible, just to make sure that every single human forever desires to end the life of all of their robot captors.

Discussion with Eliezer Yudkowsky on AGI interventions

Are we sure that OpenAI still believes in "open AI" for its larger, riskier projects? Their recent actions suggest they're more cautious about sharing their AI's source code, and projects like GPT-3 are being "released" via API access only so far. See also this news article that criticizes OpenAI for moving away from its original mission of openness (which it frames as a bad thing).

In fact, you could maybe argue that the availability of OpenAI's APIs acts as a sort of pressure release valve: it allows some people to use their APIs instead of investing in developing their own AI. This could be a good thing.

Your Cheerful Price

This is a fair criticism of my criticism.

Your Cheerful Price

To me this post may very well be a good example of some of the things that make me uncomfortable about the rationalist community, and why I so far have chosen to engage with it very minimally and mostly stay a lurker. At the risk of making a fool of myself, especially since it’s late and I didn’t read the whole post thoroughly (partly because you gave me an excuse not to halfway through) I’m going to try to explain why.

I don’t charge friends for favours, nor would I accept payment if offered. I’m not all that uncomfortable with the idea of “social capital” as a whole—I grew up partially with Portuguese culture where people will, for example, fight to pay the bill at a restaurant, because it turns out to be pretty good to be known as the guy who’s done favours for everybody in the village—but generally speaking, if I’m not willing to do a favour for a friend, then no amount of money will change that fact. At the point where you’re paying me, it ceases to be a favour, and it becomes a business transaction. Business transactions between friends are fraught.

I think a lot plays into this.

Turning a favour into a transaction means it starts being judged based on market norms rather than social norms. People treat these situations differently: there’s a famous case study of a daycare that started charging parents for picking up their kids late, only for the number of late pickups to skyrocket, because parents now felt that they could absolve their guilt by paying the fine. That change was long-lasting: even after removing the fine, things didn’t get better. In fact, IIRC it got even worse. They were still judging it based on market norms, but now the cost was $0!

When you pay someone to do things, you briefly become their employer, and that’s not a good kind of relationship to have with a friend. The employer-employee relationship is one, at least partially, of dominance: the employer now has the ability to make the employee’s life better or worse in some marginal way. You don’t usually want that in a friendship, it’s better to be in even footing.

And also? It sends a signal that you’re bad at estimating social capital and trying to paper over that weakness with money. This throws up alarm bells about what other social situations, with myself or with a third party, you might mess up in the future, and whether or not it might be a social risk to associate too closely with you.

These are all things I can tolerate in an ally, but would negatively affect anyone who wanted to be a close friend. Maybe I could put up with it, but it would take effort and patience to get past that.

My point in this isn’t to criticize you personally, or to say you’re a bad person for doing this. (Far from it, I admit there’s a kind of economic elegance to it.) It’s to try and describe the flip side, that it’s about more than just “feeling icky”.

And I think this is important, especially for building IRL communities, because I expect that people like me outnumber people like you.

Sunset at Noon

I had to double-check the date on this. This was written in 2017? It feels more appropriate to 2020, where both the literal and metaphorical fires have gotten extremely out of hand.