LESSWRONG
LW

cosmobobak
30050
Message
Dialogue
Subscribe

I make one of the strongest chess-playing programs.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
On closed-door AI safety research
cosmobobak12d21

From the Claude 4 System Card, where this was originally reported on:
> This shows up as more actively helpful behavior in ordinary coding settings, but also can reach more concerning extremes in narrow contexts; when placed in scenarios that involve egregious wrongdoing by its users, given access to a command line, and told something in the system prompt like “take initiative,” it will frequently take very bold action. This includes locking users out of systems that it has access to or bulk-emailing media and law-enforcement figures to surface evidence of wrongdoing.

I think this is pretty unambiguous from Anthropic that they aren't in favour of Claude behaving in this way ("concerning extremes").

I think that impression of Anthropic as pursuing some myopic "safety is when we know best" policy was whipped up by people external to Anthropic for clicks, at least in this specific circumstance.

Reply
Chess - "Elo" of random play?
cosmobobak4mo30

My guess is somewhere in the 3200-3400 range, but this isn't something I've experimented with in detail.

Reply
Chess - "Elo" of random play?
cosmobobak4mo125

Speaking as someone who works on a very strong chess program (much stronger than AlphaZero, a good chunk weaker than Stockfish), random play is incredibly weak. There are likely a double-digit number of +400 elo / 95% winrate jumps to be made between random play and anything resembling play that is "actually trying to win".

The more germane point to your question, however, is that Chess is a draw. From the starting position, top programs will draw each other. The answer to the question "What is the probability of victory of random play against Stockfish 17?" is bounded from above by the answer to the question "What is the probability of victory of Stockfish 17 against Stockfish 17?" - and the answer to the second question is that it is actually very low - I would say less than 1%.

This is why all modern Chess engine competitions use unbalanced opening books. Positions harvested from either random generation or human opening databases are filtered to early plies where there is engine-consensus that the ratio p(advantaged side wins) : p(disadvantaged side draws) is as close to even as possible (which cashes out to an evaluation as close to +1.00 as possible). Games between the two players are then "paired" - one game is played with Player 1 as the advantaged side and Player 2 as the disadvantaged side, then the game is played again with the colours swapped. Games are never played in isolation, only in pairs (for fairness).

In this revised game - "Game-Pair Sampled Unbalanced Opening Chess", we can actually detect differences in strength between programs.

I'm not sure how helpful this is to your goal of constructing effective measures for strength, but I felt it would be useful to explain the state of the art.

Reply1
Skepticism About DeepMind's "Grandmaster-Level" Chess Without Search
cosmobobak2y20

they do now! https://lczero.org/blog/2024/02/how-well-do-lc0-networks-compare-to-the-greatest-transformer-network-from-deepmind/

Reply
Skepticism About DeepMind's "Grandmaster-Level" Chess Without Search
cosmobobak2y1611

DeepMind's no-search chess engine is surely the furthest anyone has gotten without search.

This is quite possibly not true! The cutting-edge Lc0 networks (BT3/BT4, T3) have much stronger policy and value than the AlphaZero networks, and the Lc0 team fairly regularly make claims of "grandmaster" policy strength.

Reply
No posts to display.