joseph_c

Wiki Contributions

Comments

Sorted by

I recently came across Backpack Language Models and wanted to share it in case any AI interpretability people have not seen it. (I have yet to see this posted on LessWrong.)

The main difference between a backpack model and an LLM is that it enforces a much stricter rule to map inputs' embeddings to output logits. Most LLMs allow the output logits to be an arbitrary function of the inputs' embeddings; a backpack model requires the output logits to be a linear transformation of a linear combination of the input embeddings. The weights for this linear combination are parameterized by a transformer.

The nice thing about backpack models is that they are somewhat easier to interpret/edit/control: The output logits are a linear combination of the inputs' embeddings, so you can directly observe how changing the embeddings changes the outputs.

joseph_c2-1

Most students, 48 percent, claimed to be Native American on their application....

According to Intelligent.com Managing Editor Kristen Scatton, the prevalence of applicants who claim Native American ancestry is possibly due to the popular narrative that for many Americans, a small percentage of their DNA comes from a Native American tribe.

 

Maybe these students are purposely misinterpreting "Native American" to be someone who was born and raised in the United States, perhaps with ancestors born and raised in the US as well. This is actually the older sense of the term "Native American", found, for example, in the name of the Native American Party back in the mid-1800s.

includeIt is written in More Dakka:

If something is a good idea, you need a reason to not try doing more of it.

Taken at face value, it implies the contrapositive:

If something is a bad idea, you need a reason to not try doing less of it.

 

This is not the contrapositive. It is not even the opposite.

Parfit's Hitchhiker: You're stranded in the desert and Omega comes up. It will give you a ride out of the desert iff it predicts you'd give it 10,000 dollars upon reaching civilization again. You get a ride. When in civilization again, do you go over to the bank and withdraw some money? Well, policies which pay up in this specific situation get (value of a life - 10,000 dollars) more than policies which don't pay in this specific situation, which just die.

Why is this called Parfit's Hitchhiker? Who is the Parfit it is referring to? Where was this scenario first written up? (I'm trying to dig up the original reference.)

Not for Mormons. They don't believe in an omnipresent God.

joseph_c6247

Well, what are your actual steps? Or is this just advertisement?

Do you still live in Utah?

Did your family cut you off?

Do you know about [r/exmormon](https://old.reddit.com/r/exmormon/)?

Maybe try controlling for age? I think young people are both less likely to have signed up for cryonics (because they have less money and are less likely to die) and also have higher probabilities of cryonics working for them (because cryonics will improve by the time they need it).

joseph_c139

 

This graph seems to match the rise of the internet. Here's my alternate hypothesis: Most people are irrational, and now it's more reasonable to call them crazy/stupid/fools because they have much greater access to knowledge that they are refusing/unable to learn from. I think people are just about as empathetic as they used to be, but incorrect people are less reasonable in their beliefs.

The trick here is that both equations contain  which is the hardest to calculate, and that number drops out when we divide the equations.

 

You have a couple typos here. The first centered equation should not have a $P(\bar H H | X)$ but instead have $P(\bar H | X)$, and the inline expression should be $P(D | X)$, not $P(D | H)$.

Load More