drgunn

Communications consultant for AI safety and policy.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
drgunn10

I'm really glad to see taste enter the discourse! Taste is a meta-rational skill and meta-rationality - the art of relevant realization, how to not get stuck optimizing up to a local maximum that's not actually where you want to be - has very much been lacking in public conversation about how to get more of the good AI can bring while also doing common-sense things to manage risk.

Why have I been concerned about this? Optimizing up to a local maximum which isn't actually where you want to be is EXACTLY the problem that so many content platforms (Facebook, Twitter, Quora) have gotten themselves into and my impression has been that the people building AI products are drawn from the same population. So seeing taste get discussed is very welcome and, if this continues, will definitely prompt a positive update to my expectation of how things are going to turn out.

drgunn10

A useful direction for further research would be to analyze fiction and non-fiction as separate corpora. Longer sentences may reflect a tendency towards hypotaxis, and they may also be a deliberate stylistic choice for fiction writers, perhaps setting a more languid and relaxed tone not appropriate for much non-fiction.

drgunn21

Fanelli is a good, if dated reference for this. Another important point is that there are levels of misconduct in research, ranging from bad authorship practices to outright fabrication of results, with the less severe practices being relatively more common: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4269469/

Aside from all that, there's irreproducibility, which doesn't arise from any kind of deliberate misconduct, but still pollutes the epistemic commons: https://www.cos.io/rpcb

drgunn40

As someone with experience in BSL-3 labs, BSL feels like a good metaphor to me. The big issue with the RSP proposal is that it's still just a set of voluntary commitments that could undermine progress on real risk management by giving policymakers a way to make it look like they've done something without really doing anything. It would be much better with input from risk management professionals.

drgunn10

I'm confused.The protest is now listed at 4. Have you coordinated with them?

drgunn30

I like it, but it feels like you could have worked snakes in there somehow: https://www.vectorsofmind.com/p/the-snake-cult-of-consciousness

drgunn17-2

X-risk discussions aren't immune from the "grab the mic" dynamics that affect every other cause advocacy community.

There will continue to be tactics such as "X distracts from Y" and "if you really cared about X you would ..." unless and until people who care about the cause for the cause's sake can identify and exclude those who care about the cause for the sake of the cultural and social capital that can be extracted. Inclusivity has such a positive affect halo around it that it's hard to do this, but it's really the only way.

Longer-form of the argument: https://meaningness.com/geeks-mops-sociopaths

drgunn10

When I hear someone saying,

"Because it is asking to me to believe it completely if I believe it at all, I feel more comfortable choosing to consider it “false” on my terms, which is merely that it gave me no other choice because it defined itself to be false when it is only believed weakly."

what I think is "of course there are strong and weak beliefs!" but true and false is only defined relative to who is asking and why (in some cases), so you need to consider the context in which you're applying LoEM.

In other words, LoEM applies to "Does 2+2=4?" but it does not apply to "Is there water in the fridge?", unless the context is specified more carefully.

It's obviously an error to only have 100% or 0% as truth values for all propositions, and it's perhaps less obviously an error to have the same probabilities that a proposition is true across all possible contexts in which that proposition might be evaluated.

More here: https://metarationality.com/refrigerator

drgunn20

You seem to be talking about "combinatorial explosion". It's a classic problem in AI, and I like John Vervaeke's approach to explaining how humans solve the problem for themselves. See: http://sites.utoronto.ca/jvcourses/jolc.pdf

No one has solved it for AI yet.

drgunn10

It's a very interesting model of the world that tastes 18 different brands of store-bought sauce and doesn't compare it to just making your own. Add "seriouseats" to any recipe-related query and you'll get recipes that both work and taste like they're supposed to. They may eventually fire their writers and replace them with AI trained on blogspam recipes, so exploit this knowledge while you can. Surprisingly few people know about it, given how much utility basic culinary knowledge can add to your life.

Load More