LESSWRONG
LW

AI
Personal Blog

26

ARENA 6.0 - Call for Applicants

by JamesH, JScriven, David Quarel, CallumMcDougall, James Fox
4th Jun 2025
7 min read
3

26

AI
Personal Blog

26

ARENA 6.0 - Call for Applicants
2Sheikh Abdur Raheem Ali
1JamesH
1Sheikh Abdur Raheem Ali
New Comment
3 comments, sorted by
top scoring
Click to highlight new comments since: Today at 4:23 PM
[-]Sheikh Abdur Raheem Ali3mo20

I decided to put an application in. Elapsed time to fill this out was 6 hours— from 9 PM to 3 AM local time. Only one data point, but I’m probably slower than median by a fair margin.

Reply
[-]JamesH3mo10

I'm sorry it took you so long to fill out! I hope you're correct that you were slower than the median by a decent amount, since I don't want the application to take up quite so much of people's time. However, definitely appreciate you letting us know how long it took you.

We want to try and get a sense of people's experience in AI safety, future career plans, and see how they engage with AI safety material through the application (as well as their technical experience, since it can be a pretty technically demanding course), as well as a bunch of logistical stuff, of course.

I find that we tend to get a lot of signal from virtually all the parts of the application (apart from some of the logistical stuff, but I imagine that stuff is relatively quick to fill out). We have thought about trying to cut it down somewhat, but found it difficult to remove anything.

Reply
[-]Sheikh Abdur Raheem Ali3mo10

I believe that the ARENA application process is fine, and I definitely didn't mean to imply that taking this much time was due to the form being bloated. I did not predict that it would take me six hours, but I am also not surprised that it took that long.

It may be helpful to outline a few factors that could have played a role here, mostly ones obvious and known to my inside-vew model, though I'll err on the side of providing more detail than strictly necessary.

  • Health:
    • ADHD: I take stimulant medication to compensate for this but the effects had worn off.
    • Stress: I was supposed to travel, but recent UAE airspace closures had cancelled my flight.
    • Exhaustion: It was the end of the day, so I was tired and thinking more slowly than usual.
    • Allergies: I was sick and took antihistamines to prevent being interrupted by sneezes.
  • Competitiveness:
    • Since admissions are holistic, spending a lot more time on the early stages of the application help me eke out a small advantage relative to other candidates who will do better than me on later stages of the process.
    • I am happy to invest a lot of resources into increasing the marginal probability doing ARENA as I believe the expected return would be significant.
  • Lack of Practice:
    • I'm quite selective when it comes to targeting opportunities and only apply to a role when I'd be excited to accept an offer, so don't have a lot of experience filling out applications.
    • In general, I am measurably slower at all forms of non-LLM assisted writing compared to peers and this is a fundamental weakness where I need to work on fixing errors in cognitive algorithms to improve.

Finally, I think that a 4x difference in allocated vs consumed time to complete a task is worth digging into further, so I'll go over the screen recording of that interval to understand what happened in this case. The following breakdown of where the time went would make it clearer whether we can expect others to have also taken extra steps not accounted for in the original 1.5 hr estimate.

Planned timeline:

9:00 PM Start application.

10:30 PM Finish and submit.

Actual Timeline:

9:00 PM: Wrap-up what I was doing before.

9:07 PM: Form opened

9:08 PM: Start "Career Plans" question.

9:59 PM: Handle interrupt.

10:01 PM: Return to application.

10:21 PM: Finish "Career Plans" question.

10:22 PM: Ask for feedback in family group.

10:24 PM: Save paragraph for future use.

10:25 PM: Start "Why ARENA" question.

10:30 PM: Deadline exceeded, triggering forced context shift[1]

10:31 PM: Decide to prioritize completing ARENA application over staying on schedule.

11:20 PM: Finish paragraph 1/3 of answer "Why ARENA".

11:29 PM: Look up (scope: ARENA material [extended][2]) to verify a claim.

11:33 PM: Complete fact-check, return to writing.

11:41 PM: Look up (scope: Gemini developer documentation) to verify a claim.

11:43 PM: Complete fact-check, return to writing.

11:57 PM: Finish paragraph 2/3 of answer "Why ARENA".

12:03 AM: Look up (scope: explorables, neuronpedia, gmail, DMs) to verify a claim.

12:09 AM: Complete fact-check, return to writing.

12:23 AM: Finish paragraph 3/3 of answer "Why ARENA".

12:24 AM: Ask for feedback in family group.

12:25 AM: Start "Logistics" section.[3]

12:26 AM: Complete "Logistics" section.

12:27 AM: Double-check "Logistics" section.[4]

12:28 AM: Ask for feedback in family group.

12:29 AM: Read questions in "AI Safety Experience" and "Technical experience" sections.

12:30 AM: Look up (scope: "How to work through the ARENA program on your own" by Leon Lang)[5]

12:31 AM: Complete fact-check, return to writing.

12:32 AM: Start "Mentor recommendation" question.

12:33 AM: Finish answer for "Mentor recommendation" question.

12:34 AM: Paste answer to "Tell me about your experience in AI Safety" question from notes.

12:35 AM: Write answer to "Tell us about your most impressive technical accomplishment" question.

12:36 AM: Delete partial answer.[6]

12:37 AM: Start "Tell me about your coding/ML experience" question.

12:38 AM: Look up (scope: git historical activity statistics) to verify a claim.

12:40 AM: Abort query, return to writing.

12:45 AM: Finish answer for "Tell me about your coding/ML experience" question.[7]

12:46 AM: Ask for feedback in family group.

12:49 AM: Think about past technical accomplishments I've had.

12:52 AM: Select one from my resume to write about.

1:06 AM: Finish answer for "Tell me about your most impressive technical accomplishment"[8]

1:07 AM: Ask for feedback in family group.

1:09 AM: Start editing my standard resume template to customize it for ARENA[9]

1:17 AM: "This shouldn't take more than 20 minutes total", start timer for completing task.[10]

1:29 AM: Deadline exceeded, triggering forced context shift.[11]

1:30 AM: Upload resume

1:31 AM: Start "Technical Alignment Research Agenda" question

1:32 AM: Look up (scope: turntrout.com, ninapanickssery.com) to verify a claim.

1:35 AM: Complete fact-check, return to writing.

1:51 AM: Look up (scope: "Distillation robustifies unlearning" by Team Shard) to verify claim.

2:14 AM: Finish answer for "Technical Alignment Research Agenda" question[12]

2:15 AM: Ask for feedback in family group.

2:16 AM: Start "Read and Understand Alignment Faking summary" question.

2:17 AM: Look up (scope: AF extension experiments, proposal document, overleaf with paper draft)

2:26 AM: Complete fact-check, return to writing.[13]

2:47 AM: Finish answer for "Read and Understand Alignment Faking summary" question.

2:48 AM: Ask for feedback in family group.

2:49 AM: Final editing pass for typos before submission.

2:50 AM: Submit ARENA 6.0 application.

2:52 AM: Update weekend plans to account for circadian rhythm disruption[14]

3:08 AM: Post comment here to report actual time taken for application.

 

  1. ^

    Using a browser extension https://addons.mozilla.org/en-GB/firefox/addon/tab-scheduler-auto-open-close/ configured to automatically close current tabs and open planned tabs for next task)

  2. ^

    ARENA [core] means all materials directly included in the main branch of the ARENA repo. ARENA [extended] adds messages exported from public channels of ARENA Slack via slackdump (with verbal permission granted by Callum McDougall), saved text from crawled links to archived web articles (where the policy in robots.txt permits scraping), as well as relevant lesswrong posts such as impact reports, calls for applications, or ARENA final projects (curated manually).

  3. ^

    No save step, as it is unlikely that section "Why ARENA" can be pasted into future applications.

  4. ^

    I did this by eye since I was starting to feel pressure to speed up and didn't want to spend another look up cycle on this. In retrospect, I should have taken the few extra minutes for accuracy, since I ended up having making a mistake here and had to email a correction after.

  5. ^

    Observing that I was two hours over at this point lead to a negative update on p(accept)

  6. ^

    Realized that I had to start over since what I had written off the top of my head risked revealing potentially confidential/sensitive information that could risk violating terms of an NDA.

  7. ^

    This was really hard to write for since my instincts were strategically optimizing a myopic goal of getting into ARENA but my principles couldn't relax constraints of full truthfulness and honesty.

  8. ^

    Skipping over some ad-hoc lookups that I did to add hyperlinks to outside sources

  9. ^

    Normally I wouldn't bother, but this step was included due to feedback received during a call I'd had with a Research Manager at MATS (John Teichman).

  10. ^

    I had a lot of things opened at this point (e.g Discord, OpenReview, LinkedIn).

  11. ^

    In this case, using tex2pdf to build the in-progress project I was working on in my IDE

  12. ^

    No save step, this is well worth thinking about from scratch every time I am asked!

  13. ^

    I'd already read Scott's post when it came out so didn't need to reread it. I have been working with collaborators at Anthropic on a research project related to AF+interp but decided it may be unfair to use results from that in this application.

  14. ^

    i.e, moving a board game cafe hangout from morning to afternoon

Reply
Moderation Log
More from JamesH
View more
Curated and popular this week
3Comments

TL;DR:

We're excited to announce the sixth iteration of ARENA (Alignment Research Engineer Accelerator), a 4-5 week ML bootcamp with a focus on AI safety! Our mission is to provide talented individuals with the ML engineering skills, community, and confidence to contribute directly to technical AI safety. ARENA will be running in-person from LISA from September 1st – October 3rd (the first week is an optional review of Neural Network Fundamentals).

Apply here to participate in ARENA before 23:59 on June 21st 2025 (anywhere on Earth).

Summary:

ARENA has been successfully run five times, with alumni going on to become MATS scholars and LASR participants; AI safety engineers at Apollo Research, METR, UK AISI, and even starting their own AI safety organisations!

This iteration will run from September 1st – October 3rd (the first week is an optional review of Neural Network Fundamentals) at the London Initiative for Safe AI (LISA) in Shoreditch, London. LISA houses AI safety organisations (e.g., Apollo Research, BlueDot Impact), several other AI safety researcher development programmes (e.g., LASR Labs, PIBBSS, Pivotal, Catalyze Impact), and many individual researchers (independent and externally affiliated). Being situated at LISA brings several benefits to participants, such as productive discussions about AI safety and different agendas, allowing participants to form a better picture of what working on AI safety can look like in practice, and offering chances for research collaborations post-ARENA.

The main goals of ARENA are to:

  • Find high-quality participants;
  • Upskill these talented participants in ML skills for AI safety work;
  • Integrate participants with the existing AI safety community;
  • Accelerate participants’ career transition into AI safety.

The programme's structure will remain the same as ARENA 5.0 (see below). For more information, see our website.

Also, note that we have a Slack group designed to support the independent study of the material (join link here).

Outline of Content:

The 4-5 week programme will be structured as follows:

Chapter 0: Neural Network Fundamentals

Before getting into more advanced topics, we first cover the basics of deep learning, including basic machine learning terminology, what neural networks are, and how to train them. We will also cover some subjects we expect to be useful going forward, e.g. using GPT-3 and 4 to streamline your learning, good coding practices, and version control.

Note: Participants can optionally skip this week of the programme and join us at the start of Chapter 1 if they’re unable to attend otherwise and if we’re confident that they are already comfortable with the material in this chapter. It is recommended that participants attend, even if they’re familiar with the fundamentals of deep learning.

Topics include:

  • PyTorch basics
  • CNNs, Residual Neural Networks
  • Optimization (SGD, Adam, etc)
  • Backpropagation
  • Hyperparameter search with Weights and Biases
  • GANs & VAEs

Chapter 1 - Transformers & Interpretability

In this chapter, you will learn all about transformers and build and train your own. You'll also study LLM interpretability, a field which has been advanced by Anthropic’s Transformer Circuits sequence, and work by Neel Nanda and the GDM Interpretability Team. This chapter will also branch into areas more accurately classed as "model internals" than interpretability, for example, work on steering vectors.

Topics include:

  • GPT models (building your own GPT-2)
  • Training and sampling from transformers
  • TransformerLens
  • In-context Learning and Induction Heads
  • Indirect Object Identification
  • Superposition
  • Steering Vectors

Chapter 2 - Reinforcement Learning

In this chapter, you will learn about some of the fundamentals of RL and work with OpenAI’s Gym environment to run their own experiments.

Topics include:

  • Fundamentals of RL
  • Vanilla Policy Gradient
  • Proximal Policy Gradient
  • RLHF (& finetuning LLMs with RLHF)
  • Gym & Gymnasium environments

Chapter 3 - Model Evaluation

In this chapter, you will learn how to evaluate models. We'll take you through the process of building a multiple-choice benchmark of your own and using this to evaluate current models through UK AISI's Inspect library. We'll then move on to study LM agents: how to build them and how to elicit behaviour from them.

Topics include:

  • Constructing benchmarks for models
  • Using models to develop safety evaluations
  • Building pipelines to automate model evaluation
  • Building and eliciting LM agents

Chapter 4 - Capstone Project

We will conclude this program with a Capstone Project, where participants will receive guidance and mentorship to undertake a 1-week research project building on materials taught in this course. This should draw on the skills and knowledge that participants have developed from previous weeks and our paper replication tutorials.

Here is some sample material from the course on how to replicate the Indirect Object Identification paper (from the chapter on Transformers & Mechanistic Interpretability). An example Capstone Project might be to apply this method to interpret other circuits, or to improve the method of path patching. You can see some capstone projects from previous ARENA participants here and here.

Call for Staff

ARENA has been successful because we had some of the best in the field TA-ing with us and consulting with us on curriculum design. If you have particular expertise in topics in our curriculum and want to apply to be a TA, use this form to apply. TAs will be well compensated for their time. Please contact info@arena.education with any more questions.

FAQs:

Q: Who is this programme suitable for?

A: There’s no single profile that we look for at ARENA; in recent iterations, successful applicants have come from diverse academic and professional backgrounds. We intend to keep it this way – this diversity makes our bootcamps a more enriching learning experience for all.

When assessing applications to our programme, we like to see:

  • Applicants who genuinely care about AI safety and making the future development of AI go well;
  • Applicants who are able to code well in Python, and have some knowledge of the maths needed for modern AI (linear algebra, calculus, probability);
  • A solid understanding of how you might best contribute to technical AI safety, and how you expect ARENA to help you achieve your goals.

Since ARENA is an ML bootcamp, some level of technical skill in maths and coding will be required – more detail on this can be found in our FAQs. However, if our work resonates with you, we encourage you to apply.

Q: What will an average day in this programme look like?

At the start of the programme, most days will involve pair programming, working through structured exercises designed to cover all the essential material in a particular chapter. The purpose is to get you more familiar with the material in a hands-on way. There will also usually be a short selection of required readings designed to inform the coding exercises.

As we move through the course, some chapters will transition into more open-ended material. For example, in the Transformers and Mechanistic Interpretability chapter, after you complete the core exercises, you'll be able to choose from a large set of different exercises, covering topics as broad as model editing, superposition, circuit discovery, grokking, discovering latent knowledge, and more. In the last week, you'll choose a research paper related to the content we've covered so far & replicate its results (possibly even extend them!). There will still be TA supervision during these sections, but the goal is for you to develop your own research & implementation skills. Although we strongly encourage paper replication during this chapter, we would also be willing to support well-scoped projects if participants are excited about them.

Q: How many participants will there be?

We're expecting to accept around 30 participants in the in-person programme.

Q: Will there be prerequisite materials?

A: Yes, we will send you prerequisite reading & exercises covering material such as PyTorch, einops and some linear algebra (this will be in the form of a Colab notebook) a few weeks before the start of the programme.

Q: When is the application deadline?

A: The deadline for submitting applications is 23:59 on June 21st 2025, anywhere on Earth.

Q: What will the application process look like?

A: There will be three steps:

  1. Fill out the application form;
  2. Perform a coding assessment;
  3. Interview virtually with one of us, so we can find out more about your background and interests in this course.

Q: Can I join for some sections but not others?

A: Participants will be expected to attend the entire programme. The material is interconnected, so missing content would lead to a disjointed experience. We have limited space and, therefore, are more excited about offering spots to participants who can attend the entirety of the programme.

The exception to this is the first week, which participants can choose to opt in or out of based on their level of prior experience (although attendance is strongly recommended if possible).

Q: Will you pay stipends to participants?

A: We won't pay stipends to participants. However, we will be providing housing and travel assistance to in-person participants (see below).

Q: Which costs will you be covering for the in-person programme?

A: We will cover all reasonable travel expenses to and from London (which will vary depending on where the participant is from) and visa assistance, where needed. Accommodation, meals, and drinks and snacks will also all be included.

Q: I'm interested in trialling some of the material or recommending material to be added. Is there a way I can do this?

A: If either of these is the case, please feel free to reach out directly via an email to info@arena.education (alternatively, send us a LessWrong/EAForum message). We'd love to hear from you!

Links to Apply:

Here is the link to apply as a participant. You should spend no more than 90 minutes on it.

Here is the link to apply as a TA. You shouldn't spend longer than 30 minutes on it.

We look forward to receiving your application!