LESSWRONG
LW

Arturs
0120
Message
Dialogue
Subscribe

I am a Technical AI Governance researcher with interests in animal ethics, multilingual AI capabilities and safety, compute governance, and the economics of transformative AI. My background includes over 10 years of experience spanning project management, quantitative risk analysis and model validation in finance, and research in economics. I am also the founder and chair of the board at ๐˜Œ๐˜ง๐˜ง๐˜ฆ๐˜ค๐˜ต๐˜ช๐˜ท๐˜ฆ ๐˜ˆ๐˜ญ๐˜ต๐˜ณ๐˜ถ๐˜ช๐˜ด๐˜ฎ ๐˜“๐˜ข๐˜ต๐˜ท๐˜ช๐˜ข and a board member of the animal advocacy organization ๐˜‹๐˜ป๐˜ชฬ„๐˜ท๐˜ฏ๐˜ช๐˜ฆ๐˜ฌ๐˜ถ ๐˜ฃ๐˜ณ๐˜ชฬ„๐˜ท๐˜ชฬ„๐˜ฃ๐˜ข.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
Access to powerful AI might make computer security radically easier
Arturs1y1-1

These frontier models could still be vulnerable to stealth (e.g. โ€œsleeper agentโ€) attacks, specialist models, and stealth attacks by specialist models. The balance depends on the ability gap โ€“ if the top model is way ahead of others, then maybe defence dominates attack efforts. But a big ability gap does not seem to be playing out, instead there are several frontier models near-frontier, and lots of (more or less) open source stuff not far behind.

Reply
Predictable updating about AI risk
Arturs2y*10

Seems like a stark case of contrast between Bayesianism and the way a frequentist might approach things. I.e. do not reject the null hypothesis of no significant probability until convinced by evidence, either formal arguments or by seeing real-life mishaps. Labeling something as having P(x)~0 probably helps to compartmentalize things, focus to other tasks at hand. But can lead to huge risks being neglected, like in this case of AI Alignment.

Edit: "premortem" seems like a useful exercise to align mind & gut

Reply
No wikitag contributions to display.
-1Road to AnimalHarmBench
11d
0
5The Future of Work: How Can Policymakers Prepare for AI's Impact on Labor Markets?
1y
0