LESSWRONG
LW

Legionnaire
1064320
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
2Legionnaire's Shortform
1y
10
No wikitag contributions to display.
Legionnaire's Shortform
Legionnaire3mo51

Months ago I suggested that you could manipulate the popular LLMs by mass publishing ideological text online. Well this has now been done by Russia.

Reply
So how well is Claude playing Pokémon?
Legionnaire4mo10

Me and my college educated wife recently got stuck playing Lego Star wars... Our solution was to go to Google it. Some of these games are poorly designed and very unintuitive as others have said. Especially a game this old. Seems like they should give Claude some limited Google searches at least.

The earliest Harry Potter games had help hotlines you could call, which we had to do once when I was 9.

It's hilarious it thinks the game might be broken sometimes, like an angry teenager claiming lag when he loses a firefight in CoD.

Reply
A Bear Case: My Predictions Regarding AI Progress
Legionnaire4mo74

It will not meaningfully generalize beyond domains with easy verification

Why can't we make every domain have automated verification? (I wont claim easy, but easy enough to do with finite resources) Agency, for instance, is verifiable in competitive games of arbitrary difficulty and scale. Just check who won. DeepMind has already done this to some degree with language models and virtual agents a year ago. https://deepmind.google/discover/blog/sima-generalist-ai-agent-for-3d-virtual-environments/

Every other trait we care about is instrumental in agency to some degree, and the games can be customized to focus on various aspects as well, just like you focus a class in school.

Reply
Have LLMs Generated Novel Insights?
Answer by LegionnaireFeb 25, 202572

It's hard to see what a novel insight is exactly. Any example can be argued against. Can you give an example of one? Or of one you've personally had?

Various LLMs can spot issues in code bases that are not public. Do all of these count?

Reply
Legionnaire's Shortform
Legionnaire6mo20

Well that puts my concern to rest. Thanks!

Reply
Numberwang: LLMs Doing Autonomous Research, and a Call for Input
Legionnaire6mo30

Would also love to take the tests. If possible you could grab human test subjects from certain areas: a less wrong group, a reddit group, etc.

Reply1
Legionnaire's Shortform
Legionnaire6mo11

Who is aligning lesswrong? As lesswrong becomes more popularized due to AI growth, I'm concerned the quality of lesswrong discussion and posts has decreased since creating and posting have no filter. Obviously no filter has been a benefit while lesswrong was a hidden gem, only visible to those who can see its value. But as it becomes more popular, i think it should be obvious this site would drop in value if it trended towards reddit. Ideally existing users prevent that, but obviously that will tend to drift if new users can just show up. Are there methods in place for this issue?

Specific example: lots of posts seem like rehashes of things that have already been plainly discussed, and the quick takes section, and discussion on Discord, do a great job of cutting down on this particular issue. So maintaining high quality posts is not a pipe dream!

Reply
Yoav Ravid's Shortform
Legionnaire10mo70

LLMs can be very good at coming up with names with some work:

A few I liked:
Sacrificial Contest
Mutual Ruin Game
Sacrificial Spiral
Universal Loss Competition
Collective Sacrifice Trap
Competition Deadlock
Competition Spiral
Competition Stalemate
Destructive Contest
Destructive Feedback Competition 
Conflict Feedback Spiral
 

Reply1
Legionnaire's Shortform
[+]Legionnaire10mo-6-2
Legionnaire's Shortform
Legionnaire1y10

Speculation: LLM Self Play into General Agent?
Suppose you got a copy of GPT4 post fine tuning + hardware to train it. How would the following play out?
1. Give it the rules and state of a competitive game, such as automatically generated tic-tac-toe variants.
2. Prompt it to use chain of thought to consider the best next move and select it.
3. Provide it with the valid set of output choices (like a json format determining action and position, similar to AutoGPT)
4. Run two of these against each other continuously, training on the results of the victor which can be objectively measured by the game's rules.
5. Benchmark it against a tiny subset of those variants that you want to manually program a bot with known ELO / have a human evaluate it.
6. Increase the complexity of the game when it reaches some general ability (eg tic tac toe variants > chess variants > Civilization 5 The Videogame variants) 

Note this is similar to what Gato did. https://deepmind.google/discover/blog/a-generalist-agent/

This would have an interesting side effect of making its output more legible in some ways than a normal NN agent, though I suppose there's no guarantee the chain of thought would stay legible English unless additional mechanisms were put in place, but this is just a high level idea.

Reply
Load More
2Legionnaire's Shortform
1y
10
3Making 2023 ACX Prediction Results Public
Q
1y
Q
9
5The Moral Copernican Principle
2y
7
40Why will AI be dangerous?
3y
13