Dark Matters

I think what's happening is basically that the pink shows where the visible mass is, but the purple shows where the mass should be according to gravitational lensing. Dark matter should pass straight through, and that is what we see according to lensing, even though the pink lags behind because it can collide (since it's mostly the hot plasma).

At least, I think that's what's happening... I myself am really confused and am pretty unconfident in that explanation.

I'm also confused as to what modified gravity predicts, and how bullet clusters disprove it. I guess what we'd see is that modified gravity would alter the gravity around the visible mass, not just make it magically act like it just passed through. Ie, a lot about gravity would have to change for such a drastic difference between the mass as perceived through x-rays and the mass as perceived through gravitational lensing.

Toward A Bayesian Theory Of Willpower

This reminds me of lukeprog's post on motivation (I can't seem to find it though...; this should suffice). Your model and TMT kind of describe the same thing: you have more willpower for things that are high value and high expectancy. And the impulsiveness factor is similar to how there are different "kinds" of evidence: eg you are more "impulsive" to play video games or not move, even though they aren't high value+expectation logically.

What are your greatest one-shot life improvements?

What does he do specifically? It's very unclear just from reading the Amazon description. Or is it like an entire program. I'm skeptical: I have never heard of this anywhere else, so it seems like one of those $100-bill-on-the-subway-floor type things.

What are your greatest one-shot life improvements?

it needed me to also have an active project I was working on that I actually enjoyed. I think otherwise I would have found other ways to distract myself and eventually undermined it to the point that I gave up.

Same with me. Although it's still better than nothing: the usual distractions are more habit than actually fun, and I've found that I read more interesting things instead of just mindlessly browsing social media.

What are your greatest one-shot life improvements?

I like to use the add-on LeechBlockNG (I don't know if you can use it on mobile). You can use it to outright block sites, but also delay access to the site before you actually enter, and also put time limits. The delay is something I haven't seen other apps/add-ons use a similar feature, and it's kind of a deal-breaker, since it solves the problem of "instant gratification" that makes social media (etc.) addicting.

How to Ignore Your Emotions (while also thinking you're awesome at emotions)

I really agree with this. I have been thinking that we should "default to privacy", because if we think we have to share it, we will change our thoughts because of the social anxieties/pressures. (It's similar to that experiment that demonstrated people make better decisions if they didn't have to come to a solution first (I just remember this from reading HP:MOR).) Only after we reach the answer, (socially) unbiased, then we can decide to share it.

I don't think privacy means dishonesty. I personally really dislike lying, and I think it's because acting with false information sort of takes away their free will, and more practically, this creates a lot of uncertainty. But I think you can be honest about how you withhold information, to an extent: instead of lying, you can just say, "I won't tell you" or something like that. (I'm not sure how much that is based on the practicality of it and how much is it is a like a terminal value.)

I'm sort of confused by radical honesty. Is it really, truly, "radical"? Literally everyone has intrusive thoughts, and I personally sometimes have intrusive thoughts about raping or killing or saying racial slurs. I guess that's just a nitpick, because I can easily see how to be "maximally" honest (compared to normal communication).

Things you are supposed to like

I feel like I feel a similar thing, but with regards to effective altruism and learning intellectual things. I sometimes ask myself, "are my beliefs around EA and utilitarianism just 'signaling'", especially since I'm only in high school and don't really have any immediate plans. But I'm also not a very social person, and when I do talk to others I don't usually talk about EA. I guess I'm not a very conscientious person: I like the idea of "maximizing utility" and learning cool things, but my day-to-day fun things (outside of "addictions": social media, games) are just like reading books and essays and listening to music. It's as if, I "want to like" EA and learning. Like, I don't really see any point in being rich and famous, unless you're going to do good with it, and so just doing to minimum to enjoy my life and not be a net negative in utils seems fine. (That is usually a thought I have when I'm sad+unmotivated.) Does it even make sense to say that EA/utilitarianism is just signaling? Is there any reason for me to make my actions inline with my "true" preferences (i.e. egoism)? As in: if it wasn't, "should" I listen to my "true" preferences? grouchymusicologist seems to think, no: whether or not we get those preferences from repetition and socialization, they're still our preferences.

jp's Shortform

Why don't you just use MathJax? Maybe this wasn't the case when you wrote this comment, but there should be a button that just applies the formatting, and Ankidroid can render it.

Load More