(no title)
joshhart | 1 year ago
That said the conclusion that it's a good model for cheap is true. I just would be hesitant to say it's a great model.
joshhart | 1 year ago
That said the conclusion that it's a good model for cheap is true. I just would be hesitant to say it's a great model.
A_D_E_P_T|1 year ago
What's more, DeepSeek doesn't seem capable of handling image uploads. I got an error every time. ("No text extracted from attachment.") It claims to be able to handle images, but it's just not working for me.
When it comes to math, the two seem roughly equivalent.
DeepSeek is, however, politically neutral in an interesting way. Whereas GPT-4o will take strong moral stances, DeepSeek is an impressively blank tool that seems to have no strong opinions of its own. I tested them both on a 1910 article critiquing women's suffrage, asking for a review of the article and a rewritten modernized version; GPT-4o recoiled, DeepSeek treated the task as business as usual.
tkgally|1 year ago
Have you tried asking it about Tibetan sovereignty, the Tiananmen massacre, or the role of the communist party in Chinese society? Chinese models I've tested have had quite strong opinions about such questions.
theanonymousone|1 year ago
mvdtnz|1 year ago
jchook|1 year ago
On HumanEval, I see 90.2 for GPT-4o and 89.0 for DeepSeek v2.5.
- https://blog.getbind.co/2024/09/19/deepseek-2-5-how-does-it-...
- https://paperswithcode.com/sota/code-generation-on-humaneval
selfhoster11|1 year ago
Having used the full GPT-4, GPT-4 Turbo and GPT-4o for text-only tasks, my experience is that this is roughly the order of their capability from most to least capable. In image capabilities, it’s a different story - GPT-4o unquestionably wins there. Not every task is an image task, though.
stefan_|1 year ago
selfhoster11|1 year ago
GaggiX|1 year ago