They just recently released the r1-0528 model which was a massive upgrade over the original R1 and is roughly on par with the current best proprietary western models. Let them take their time on R2.
At this point the only models I use are o3/o3-pro and R1-0528. The OpenAI model is better at handling data and drawing inferences, whereas the DeepSeek model is better at handling text as a thing in itself -- i.e. for all writing and editing tasks.
With this combo, I have no reason to use Claude/Gemini for anything.
People don't realize how good the new Deepseek model is.
My experience with R1-0528 for python code generation was awful. But I was using a context length of 100k tokens, so that might be why. It scores decently in the lmarena code leaderboard, where context length is short.
A lemon is on-par with the best western models for the majority of use cases because they do not require "state of the art" intelligence to solve or respond to the user's query. This is what the benchmarks show.
For anything that requires "AI level of intelligence", the difference is vast.
A_D_E_P_T|8 months ago
With this combo, I have no reason to use Claude/Gemini for anything.
People don't realize how good the new Deepseek model is.
energy123|8 months ago
Art9681|8 months ago
For anything that requires "AI level of intelligence", the difference is vast.