Some tasks overlap with what I would want a hypothetical smart human assistant to do: Implement ML experiments and interfaces. Read over my hundreds of pages of drafts, connect ideas to relevant prior work, formalize what makes sense to be formalized and derive implications within the formalism, suggest and perform experiments to test hypotheses, write the ideas and findings up into legible posts. Summarize conversations and meetings. Brainstorm and roleplay useful simulacra with me.
However, I do not think that an Assistant character is the best or only interface AI can give us re augmenting alignment research. I want a neocortex prosthesis that has a more powerful imagination than I, that knows vastly more, is better at math, writing, critical thinking, programming, etc, and which I can weave my thoughts and context into with high bandwidth and minimal overhead, and which is retargetable to any intention I might have. Oh, and which can instantiate Assistants or any other simulacra that might come in handy for the situation.
Sorry if this isn't as specific as you asked for; there are several reasons I didn't describe e.g. the ML experiments I'd like an assistant to do more specifically, mostly laziness.
Also, if you haven't yet, you should check out Results from a survey on tool use and workflows in alignment research.
I disagree with the assumption that AI is "narrow". In a way GPT is more generally intelligent than humans, because of the breadth of knowledge and type of outputs, and it's actually humans who outperform AI (by a lot) at certain narrow tasks.
And an assistance can include more than asking a question and receiving an answer. It can be exploratory with the right interface to a language model.
(Actually my stories are almost always exploratory, where I try random stuff, change the prompt a little, and recursively play around like that, to see what the AI will come up with)
Related to the above: in my opinion thinking of specific tools is the wrong framing. Like how a gun is not a tool to kill a specific person, it kills whoever you point it at. And a language model completes whichever thought or idea you start, effectively reducing the time you need to think.
So the most specific I can get is I'd make it help me build tooling (and I already have). And the better the tooling the more "power" the AI can give you (as George Hotz might put it).