## LESSWRONGLW

Michaël Trazzi

Big picture of phasic dopamine

Right I just googled Marblestone and so you're approaching it with the dopamine side and not the acetylcholine. Without debating about words, their neuroscience paper is still at least trying to model the phasic dopamine signal as some RPE & the prefrontal network as an LSTM (IIRC), which is not acetylcholine based. I haven't read in detail this post & the one linked, I'll comment again when I do, thanks!

Big picture of phasic dopamine

Awesome post! I happen to also have tried to distill links between RPE and phasic dopamine in the "Prefrontal Cortex as a Meta-RL System" of this blog.

In particular I reference this paper on DL in the brain & this other one for RL in the brain. Also, I feel like the part 3 about links between RL and neuro of the RL book is a great resource for this.

Curated conversations with brilliant rationalists

for reference of how costly transcripts are, the first "speech-to-audio" conversion is about $1.25 per minute, and it could take 1x the time of the audio to fix the mistakes when both have native accents, and up to 2x the audio time for non-native speakers. For a 1h podcast, this would amount to$75 + hourly rate, so roughly $100/podcast. Additionally, there's a YT-generated-subtitles free alternative. I'm currently trying this out, I'll edit this to let you know how long it takes to fix them per audio hour. Curated conversations with brilliant rationalists great idea! blue yeti used to be a relatively cost-effective option ($100) for US/Canada. For Europe, I'd recommend the t.bone which comes with a suitcase, pop filter and support for $70 (including shipping). for headsets I'd recommend any studio one for about$50, such as the Audio Technica ones.

Announcing The Inside View Podcast

Ace I'll try that too, thanks!

Announcing The Inside View Podcast

Thanks for all of those tips. I'll definitely try rev!

Announcing The Inside View Podcast

done! should be live in a few hours

Announcing The Inside View Podcast

hey, dunno why it's obfuscated, here it is! https://anchor.fm/s/56df2194/podcast/rss

Announcing The Inside View Podcast

Thanks for the feedback! I haven't really estimated how long it would take to have a transcript with speech-to-text + minor corrections,—that's definitely on the roadmap.

Re audio: cost of recording is probably like one hour (x2 if you have one guest). I think that if I were to write down the whole transcript without talking it would take me easily 4-10x the time it takes me to say it. I'm not sure on how worse the quality is though, but the way I see it conversation is essentially collaborative writing where you get immediate feedback about your flaws in reasoning. And even if I agree that a 1h podcast could be summarized in a few paragraphs, the use case is different (eg. people cooking, running, etc.) so it needs to be somewhat redundant because people are not paying attention.

Re not being interested in forecasting timelines: my current goal is to have people with different expertise share their insights on their particular field and how that could nuance our global understanding of technological progress. For instance, I had a 3h discussion with someone who did robotics competitions, and one planned with a neuroscientist student converted into a ML engineer. I'm not that interested in "forecasting timelines" as a end goal, but more interested in how to dig why people have those inside views about the future (assuming they unconsciously updated on things), so we can either destroy wrong initial reasons for believing something, or gain insight on the actual evidence behind those beliefs.

Anyway, I understand that there's a space about rigorous AI Alignment research discussions, which is currently being covered by AXRP, and the 80k podcasts also cover a lot of it, but it seems relatively low-cost to just record those conversations I would have anyway during conferences so people can decide by themselves what are the correct or bad arguments.

What will GPT-4 be incapable of?

sorry I meant a bot that played random move, not a randomly sampled go bot from KGS. agreed with GPT-4 not beating average go bot