Thanks for your thoughts here!So I do have some familiarity with the concept of deliberate practice, and I would definitely include that as part of the thing I'm talking about above. But I can also think of things that might improve a researcher's capacity that don't fall under deliberate practice.1. One researcher told me they were having frequent interruptions to their focused working time as a result of their work environment, so they made some adjustments to their work environment to prevent that. I don't think I'd call that deliberate practice, but it does seem like a big improvement.2. Motivation/procrastination. This is probably the single biggest complaint I've heard from researchers. To the extent that they find a solution to this, it probably won't end up being something in the category of "deliberate practice". It will probably look like creating incentives for themselves, or introspecting on their own preferences and motivations, or creating accountability mechanisms, etc.
If there are any alignment researchers reading this who think they would benefit from having someone to talk to about improving their research capacity, I’m happy to be that person.
I’m offering free debugging-style conversations about improving research capacity to any alignment researchers who want them. Here’s my calendly link if you’d like to grab time on my calendar: https://calendly.com/dcjones15/60min .
I’m not claiming to have any answers or ready made solutions. I primarily add value by asking questions to elicit your own thoughts and help you come up with your own improvement plans that address your specific needs. A number of researchers have told me these conversations are productive for them, so the same may be true for you.
I'm not new to reading LessWrong, but I am new to posting or commenting here. I plan to be more active in the future. I care about the cause of AI Alignment, and am currently in the process of shifting my career from low-level operations work at MIRI to something I think may be more impactful: I.e. supporting alignment researchers in their efforts to level up in research effectiveness, by offering myself as a conversational partner to help them think through their own up-leveling plans.In that spirit, here's an offer I'd like to make to any interested alignment researchers who come across this comment.The OfferFree debugging-style conversations (could be just one, or recurring) aimed at helping you become a more effective researcher. How to sign up?
Questions you may have:What would the conversation look like?
Who am I, and why might I be a good person to talk to about this?
Thanks, that helps!
These monthly threads and Stampy sound like they'll be great resources for learning about alignment research.I'd like to know about as many resources as possible for supporting and guiding my own alignment research self-study process. (And by resources, I guess I don't just mean more stuff to read; I mean organizations or individuals you can talk to for guidance on how to move forward in one's self-education). Could someone provide a link to a page that attempts to gather links to all such resources in one place? I already saw the Stampy answer to "Where Can I Learn About AI Alignment?". Is that pretty comprehensive, or are there many more resources?