(no title)
raxxor | 1 year ago
DeepSeek is awesome. Any AI task yet implemented in our business can be run from my local PC with just the smaller models. And my PC is fairly crappy to begin with.
OpenAI looks quite silly with their "we have to close everything".
manmal|1 year ago
raxxor|1 year ago
I do have some applications that process images, text and pdf files and I use smaller models for extracting embeddings. I think my system wouldn't be able to handle it with decent speed otherwise.
I do run LLM on a M1 16gb macbook air and performance is surprisingly good. Not for image synthesis though and a PC with a dedicated GPU is still significantly faster with LLM responses as well. Haven't tried to run deepseek on the macbook yet.
nejsjsjsbsb|1 year ago
Running local on CPU opens so much possibilities for smart and privacy focused home devices that serve you.
In my test it hallucinated confidently but my interest is in simple second brain like rag. "Hey thingy, what is my schedule today?"
Need it to be a bit faster though as the thinking part adds a lot of latency.
raxxor|1 year ago
It does add latency of course, but I still think that I could provide all AI needs of my company (industrial production) with a simple older off the shelf PC. My GPU is decently recent, but the smallest model of the series and otherwise the machine is a rusty bucket.
I didn't test it thoroughly yet, but I have some invoices where I need to extract info and it did a perfect job until now. But I don't think there is any LLM yet that can do that without someone checking the output.
blackeyeblitzar|1 year ago
And then there’s all the dystopian propaganda baked into these models, which threatens to misinform users at scale based on a government driven agenda. Hard to be on that team, let alone firmly, knowing that it’s giving power to a dictatorial regime.
wkat4242|1 year ago
mvc|1 year ago
And when they thought they were the only game in town, they tried to corner the market in GPUs and lock out any users who can't pony up £200/mo. Reminds me of when the likes of Oracle and IBM had companies by the balls buying bigger and bigger servers and then Google came along and showed everyone how to do horizontal scaling of cheap hardware.
raxxor|1 year ago
I haven't tested deepseek for censorship yet, but they shared their release and even their input data. And in this case you could correct its shortcomings, so propaganda would be difficult.
famouswaffles|1 year ago
The first one is definitely not true and the 2nd one is not necessarily true in the way you imagine i.e crawls of the internet will have gpt chat logs now.
timeon|1 year ago
Those models are also trained on data that was ignoring licenses / copyrighted content.