> Each model was run with the recommended hyperparameters and a maximum token limit of 64,000. No models needs more than this number of tokens
I'm a little confused by this. My assumptions (possibly incorrect!): 64k tokens per prompt, they are claiming the model wouldn't need more tokens even for reasoning
Is that right? Would be helpful to see how many tokens the models actually used.
I'm a little confused by this. My assumptions (possibly incorrect!): 64k tokens per prompt, they are claiming the model wouldn't need more tokens even for reasoning
Is that right? Would be helpful to see how many tokens the models actually used.