

Yeah, I was thinking diesel powered trains
Yeah, I was thinking diesel powered trains
This article is comparing apples to oranges here. The deepseek R1 model is a mixture of experts, reasoning model with 600 billion parameters, and the meta model is a dense 70 billion parameter model without reasoning which preforms much worse.
They should be comparing deepseek to reasoning models such as openai’s O1. They are comparable with results, but O1 cost significantly more to run. It’s impossible to know how much energy it uses because it’s a closed source model and openai doesn’t publish that information, but they charge a lot for it on their API.
Tldr: It’s a bad faith comparison. Like comparing a train to a car and complaining about how much more diesel the train used on a 3 mile trip between stations.
👏 where 👏 is 👏 his 👏 dick 👏 and 👏 prostate 👏
It’s not from David Lynch’s Dune, is it? I can’t find a scene which matches.
Edit: I’m 70% sure that this is AI generated, unless it’s a screenshot from the dune 1984 or the television series. The makeup and effects work looks too smooth, as does the feathering on the hair. Could be from a stock video platform.
Does system 76 do a bad job? I’ve looked at their prices, which are high, but haven’t heard anything bad about them.
Here is my face, sir.
Get a bidet, friend
I actually don’t think this is shocking or something that needs to be “investigated.” Other than the sketchy website that doesn’t secure user’s data, that is.
Actual child abuse / grooming happens on social media, chat services, and local churches. Not in a one on one between a user and a llm.
Yes, sorry, where I live it’s pretty normal for cars to be diesel powered. What I meant by my comparison was that a train, when measured uncritically, uses more energy to run than a car due to it’s size and behavior, but that when compared fairly, the train has obvious gains and tradeoffs.
Deepseek as a 600b model is more efficient than the 400b llama model (a more fair size comparison), because it’s a mixed experts model with less active parameters, and when run in the R1 reasoning configuration, it is probably still more efficient than a dense model of comparable intelligence.