DeepSeek launched a free, open-source large language model in late December, claiming it was developed in just two months at a cost of under $6 million.
i can also run it on my old pentium from 3 decades ago. I’d have to swap 4MiB of weights in and out constantly, it will be very very slow, but it will work.
Wth?! Like seriously.
I assume they are running the smallest version of the model?
Still, very impressive.
i can also run it on my old pentium from 3 decades ago. I’d have to swap 4MiB of weights in and out constantly, it will be very very slow, but it will work.