top | item 42832080

(no title)

bigrobinson | 1 year ago

Deepseek seems to create enormously long reasoning traces. I gave it the following for fun. It thought for a very long time (307 seconds), displaying a very long and stuttering trace before, losing confidence on the second part of the problem and getting it way wrong. GPTo1 got similarly tied in knots and took 193 seconds, getting the right order of magnitude for part 2 (0.001 inches). Gemini 2.0 Exp was much faster (it does not provide its reasoning time, but it was well under 60 second), with a linear reasoning trace, and answered both parts correctly.

I have a large, flat square that measures one mile on its side (so that it's one square mile in area). I want to place this big, flat square on the surface of the earth, with its center tangent to the surface of the earth. I have two questions about the result of this: 1. How high off the ground will the corners of the flat square be? 2. How far will a corner of the flat square be displaced laterally from the position of the corresponding corner of a one-square-mile area whose center coincides with the center of the flat area but that conforms to the surface of the earth?

discuss

order

motoboi|1 year ago

OpenAI reasoning traces are actually summarized by another model.

The reason is that you can (as we are seeing happening now) “distill” the larger model reasoning into smaller models.

Had OpenAI shown full traces in o1 answers they would have been giving gold to competition.

bigrobinson|1 year ago

That's not the point of my post, but point taken.