top | item 43852856

(no title)

wd776g5 | 10 months ago

The linked article says 14B parameters. edit and I guess the "plus" model is 21B?

discuss

order

refulgentis|10 months ago

grep "As seen above, Phi-4-mini-reasoning with 3.8B parameters outperforms models of over twice its size."

re: reasoning plus, "Phi-4-reasoning-plus builds upon Phi-4-reasoning capabilities, further trained with reinforcement learning to utilize more inference-time compute, using 1.5x more tokens than Phi-4-reasoning, to deliver higher accuracy.", presumably also 14B