

Engage your safety squints!
Engage your safety squints!
Is it a war or a cartel, though?
Not sure if OP/bot knows what this community is about… Massive shitpost, nevertheless.
Thanks for clarification!
So… as far as I understand from this thread, it’s basically a finished model (llama or qwen) which is then fine tuned using an unknown dataset? That’d explain the claimed 6M training cost, hiding the fact that the heavy lifting has been made by others (US of A’s Meta in this case). Nothing revolutionary to see here, I guess. Small improvements are nice to have, though. I wonder how their smallest models perform, are they any better than llama3.2:8b?
why are you so heavily and openly advertising Deepseek?
That article is written by DeepSeek R1 isn’t it
I sure hope everyone wiggles their fuel hose because else the golden liquid spills all over the place, posing serious biohazard. No one likes using a station with spilled liquid.
Could be the headline of an onion article