Boris Kashirin
How tokenization influences prompting?
I was thinking about how prompt differs from training data in terms of tokenization. If i am to prompt with "solution:" as opposed to "solution: " it seems like it can influence the result, as in training data last token contain some information about next token. If there is token...
GTP4 capable of limited recursive improving?
Apparently with reflection technique (answer-critique-improve) GTP4 capable of giving much better answers. But that implies it should be capable of doing essentially Alpha Go Zero type of learning! It can't do complete self play from zero as there is no ground truth for it to learn from, but that basically...
Just one data point: to me your post feel shallow in a good way. I can say: "I can't help but note how all this annoying behaviour suddenly become endearing to me when I am in love." but it feels like I am deepening your post, not contradicting it. Am I gesturing at non-terminal property you mentioned?
P.S. just noticed I am using "shallow" when there is literally "deep" in headline. Yet this is my impression.