LESSWRONG
LW

CarlShulman
13127Ω2264916820
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
peterbarnett's Shortform
CarlShulman8h159

Consider the investments in Anthropic (returning over 40x, while being a small minority of the cap table in a lab that is noticeably better than the average on AIS issues, although far from ideal) by the biggest funders of AI safety and AI policy.

If you're going to say that e.g. export controls on AI chips and fabrication equipment, non-lab AIS research (e.g. Redwood), the CAIS extinction risk letter, and similar have meaningful expected value, or are concerned about the funding limits for such activity now, I think you'll have a hard time drawing the BOTEC where e.g. Jaan shouldn't have invested. More broadly, it's possible to increase funding for those things by 100%+ while increasing capital for AI by <1%: there are trillions of dollars of market cap in AI securities (so 1% of that is tens of billions of dollars), 100%+ annualized returns have been achievable for years in those securities, and funding for all AIS-related technical and policy/political work is significantly less than <$1B/yr (and anti-AIS lobbying and spending are growing backed by returns from AI investments).

There's a big chunk of LW groupthink that seems to me to take a deontological stance against investment in AI, without apparent interest in attempting that kind of BOTEC at all, or considering that if they own index funds more than 10% of the S&P market cap is now AI (NVDA, AVGO, etc) and so they would be violating such a deontology (which I think would be a bad one to adopt).

You can take a fatalist stance that nothing can be done to help with resources, only slowing AI very slightly by personally withdrawing resources, but it will have a hard time reconciling with thinking there's any worthwhile research or advocacy to be done (and I think that stance is dangerously wrong). And you definitely shouldn't be complaining about insufficient funding for any kind of AIS work at the margin in that case.

Other things equal, I would think a doubling of, e.g. the existing OP/SFF donation mix/donation stack would be much better than the incremental AI speedup from investments generating the resources to do that, and there would be more value from different funders with different blindspots and limitations and diversification/insurance value. 

Reply
peterbarnett's Shortform
CarlShulman1d*378

To be clear, I have for many years said to people trying to earn to give on AIS related issues that investing in AI made sense, including (certain) AI labs (in certain times and conditions, externalities are much higher for early-stage labs and the character of the lab matters a lot). That does speed up AI, but the proportional change in AI speedup vs resources for AIS makes it look good when doing the BOTEC even with sizable negative effects of speedup. I think it's extremely unfortunate that some people and institutions concerned with AI safety didn't do more of that earlier, such that they could now have more than an order of magnitude more resources and could do and have done some important things with them.

E.g. I thought and said that it was reasonable for many such people to invest in early Anthropic rounds, and to do more than they did. I refrained from doing so at the time because it might create conflicts that could interfere with some of my particular advising and policy work, not because I thought in general people concerned with AIS should all avoid such investment (I thought in aggregate there was too little, not too much).

Reply41
You can, in fact, bamboozle an unaligned AI into sparing your life
CarlShulman1y70

I think if we do a poll, it will become clear that the strong majority of readers interpreted Nate's post as "If you don't solve aligment, you shouldn't expect that some LDT/simulation mumbo-jumbo will let you and your loved ones survive this" and not in the more reasonable way you are interpreting this. I certainly interpreted the post that way.

You can run the argument past a poll of LLM models of humans and show their interpretations.

I strongly agree with your second paragraph.

Reply
Partial value takeover without world takeover
CarlShulman1y*4215

If AI agents with unusual values would for a long time be mostly interested in promoting them through means other than lying in wait and taking over the world, is important because...AIs pursuing this strategy are much more visible than those hiding in wait deceptively. We might less expect AI scheming.

AIs showing bits of unintended motives in experiments or deployment would be a valuable piece of evidence re scheming risk, but such behavior would be trained against, pushing scheming behavior out towards the tails of takeover/escape with the power to resist modification. The tendency of human institutions to retrain or replace AIs to human preferences pushes towards misaligned AIs having ~0 or very high power.

Reply
Retirement Accounts and Short Timelines
CarlShulman2y161

The catastrophic error IMO is:

Five years from when you open your account there are options for taking gains out tax-free even if you're not 59.5 yet. You can take "substantially equal periodic payments", but there are also ones for various kinds of hardship.

For Roth you mostly can't take out gains tax-free. The hardship ones are limited, and SEPP doesn't let you access much of it early. The big ones of Roth conversions and just eating the 10% penalty only work for pretax.

[As an aside Roth accounts are worse for most people vs pretax for multiple reasons, e.g. pretax comes with an option of converting or withdrawing in low income years at low tax rates. More details here.]
 

In #1 if you start with $100k then it's $200k at the time you convert, and you pay $48k (24%) in taxes leaving you with $152k in your Roth 401k. It grows to $198k, you withdraw $152k and you have $46k of gains in your Roth 401k.

You pay taxes on the amount you convert, either from outside funds or withdrawals to you. If you convert $X you owe taxes on that as ordinary income, so you can convert $200k and pay $48k in taxes from outside funds. This makes pretax better.
 
Re your assumptions, they are not great for an AI-pilled saver. Someone who believes in short AI timelines should probably be investing in AI if they don't have decisive deontological objections. NVDA is up 20x in the last 5 years, OpenAI even more. On the way to a singularity AI investments will probably more than 10x again unless it's a surprise in the next few years as Daniel K argues in comments. So their 401k should be ~all earnings, and they may have a hard time staying in the low tax brackets you use (moreso at withdrawal time than contribution time) if they save a lot. The top federal tax rates are 37% for ordinary income and 23.8% for capital gains.

Paying the top federal income tax rate plus penalties means a 47% tax rate on early withdrawals from the Roth vs 23.8% from taxable. I.e. every dollar kept outside the the Roth is worth 44% more if you won't be using the account after 59.5. That's a wild difference from the standard Roth withdrawal case where there's a 0% tax rate.

A substantially larger percentage in Roth than the probability you are around to use it and care about it after 59.5 looks bad to me. From the perspective of someone expecting AI soon this advice could significantly hurt them in a way that the post obscured.


 

Reply
Retirement Accounts and Short Timelines
CarlShulman2y*2211

This post seems catastrophically wrong to me because of its use of a Roth 401k as an example, instead of a pre-tax account. Following it could create an annoying problem of locked-up funds.
 

Five years from when you open your account there are options for taking gains out tax-free even if you're not 59.5 yet. You can take "substantially equal periodic payments", but there are also ones for various kinds of hardship.

Roth earnings become tax free at 59.5. Before that, even if you use SEPP to do withdrawals without penalties you still have to pay taxes on the withdrawn earnings (some of which are your principal because of inflation). And those taxes are ordinary income rates, which top out much higher than long term capital gains tax rates. Further, the SEPP withdrawals are spaced out to reflect your whole lifetime according to actuarial tables, so if TEOTAWKI is in 10 years and the life tables have you space out your SEPP withdrawals over 40 years, then you can only access a minority of your money in that time.

For a pretax 401k where you contribute when you have a high income, the situation is very different: you get an upfront ordinary income tax deduction when you contribute, you don't get worse tax treatment by missing out on LTCG rates. And you can rollover to a Roth IRA (paying taxes on the conversion) and then access the amount converted penalty-free in 5 years (although that would trap some earnings in the Roth) or just withdraw early and pay the 10% penalty (which can be overcome by tax-free growth benefits earlier, or withdrawing in low income years).

I'm 41.5, so it's 18 years to access my Roth balances without paying ordinary taxes on the earnings (which are most of the account balances). I treat those funds as insurance against the possibility of a collapse of AI progress or blowup of other accounts, but I prefer pre-tax contributions over Roth ones now because of my expectation that probably there will be an AI capabilities explosion well before I reach 59.5. If I had all or most of my assets in Roth accounts it would be terrible.
 

Reply
Retirement Accounts and Short Timelines
CarlShulman2y40

This is pretty right for pretax individual accounts (401ks may not let you do early withdrawal until you leave), for Roth accounts that have accumulated earnings early withdrawal means paying ordinary taxes on the earnings, so you missed out on LTCG rates in addition to the 10% penalty.

Reply
Being nicer than Clippy
CarlShulman2y93

(My perennial uncertainty is: AI 1 can straightforwardly send source code / model weights / whatever to AI 2, but how can AI 1 prove to AI 2 that this file is actually its real source code / model weights / whatever? There might be a good answer, I dunno.)

They can jointly and transparently construct an AI 3 from scratch motivated to further their deal, and then visibly hand over their physical resources to it, taking turns with small amounts in iterated fashion.

AI 3 can also be given access to secrets of AI 1 and AI 2 to verify their claims without handing over sensitive data.

Reply1
[Valence series] 3. Valence & Beliefs
CarlShulman2y40

Regarding making AIs motivated to have accurate beliefs, you can make agents that do planning and RL on organizing better predictions, e.g. AIs whose only innate drives/training signal (beside short-run data modeling, as with LLM pretraining) are doing well in comprehensive forecasting tournaments/prediction markets, or implementing reasoning that scores well on various classifiers built based on habits of reasoning that drive good performance in prediction problems, even against adversarial pressures (AIs required to follow the heuristics have a harder time believing or arguing for false beliefs even when optimized to do so under the constraints).

Reply
Redirecting one’s own taxes as an effective altruism method
CarlShulman2y5749

Even if you're an anarchist who thinks taxation is theft, to say willful nonpayment of taxes to donate is effective altruism is absurd, the consequences of this are just obviously very bad, both the idea and the advocacy. One publicized case of a person willfully refusing to pay their taxes in the name of effective altruism can do much more damage to it than many such people donating a bit more, and even if a particular case is invisible, the general practice is visible (Newcomb issues). Consider how much damage SBF and FTX have done to the causes of effective altruism, pandemic prevention, AI safety. There are billions of dollars committed to effective charity, and thousands of people trying to do good effectively, and people tying commonsense wrongdoing to it with crazy rationales has a serious damaging multiplier effect on the whole.

Any dollar donated through this method is in expectation going to cost multiple dollars worth of similar donations (plausibly a huge number) equivalent through such damage. It would be much better for the world if tax scofflaws were spending their taxes due on gambling or alcohol rather than effective altruism.

Reply
Load More
90Shulman and Yudkowsky on AI progress
Ω
4y
Ω
16
43The Simulation Hypothesis Undercuts the SIA/Great Filter Doomsday Argument
Ω
4y
Ω
11
55Fractional progress estimates for AI timelines and implied resource requirements
Ω
4y
Ω
6
71Rogue AGI Embodies Valuable Intellectual Property
Ω
4y
Ω
9
18New article on in vitro iterated embryo selection
12y
34
28Why do theists, undergrads, and Less Wrongers favor one-boxing on Newcomb?
12y
300
14Normative uncertainty in Newcomb's problem
12y
32
41[Retracted] Simpson's paradox strikes again: there is no great stagnation?
13y
51
43Satire of Journal of Personality and Social Psychology's publication bias
13y
7
65Using degrees of freedom to change the past for fun and profit
14y
23
Load More