OpenAI Furious DeepSeek Might Have Stolen All the Data OpenAI Stole From Us
-
-
The way they found to train their AI cheaper was to steal it from OpenAI (not that I care). They still need GPUs to process the prompts and generate the responses.
-
It's called distilling the data, to turn the huge amount of data into a compact amount that can be used in another model.
-
-
-
-
Tamaleeeeeeeeesssssss
hot hot hot hot tamaleeeeeeeees
-
-
-
-
-
-
Right—by “take it down” I meant take down online access to their running instance of it.
-
-
You made me look ridiculously stupid and rightfully so. Actually, I take that back, I made myself look stupid and you made it obvious as it gets! Thanks for the wake up call
If I understand correctly, the model is in a way a dictionary of questions with responses, where the journey of figuring out the response is skipped. As in, the answer for the question "What's the point of existence" is "42", but it doesn't contain the thinking process that lead to this result.
If that's so, then wouldn't it be especially prone to hallucinations? I don't imagine it would respond adequately to the third "why?" in the row.
-
To add a tiny bit to what was already explained: you do actually download quite a bit of data to run it locally. The "smaller" 14b model I used was a 9GB download. The 32b one is 20GB and being all text, that's a lot of information.
-
I suspect that most usage of the model is going to be companies and individuals running their own instance of it. They have some smaller distilled models based on Llama and Qwen that can run on consumer-grade hardware.
-
Imagine if a little bit of those so many millions that so many companies are willing to throw away to the shit ai bubble was actually directed to anything useful.
-
Someone should just an make AiPU. I'm tired of all GPUs being priced exorbitantly.
-
but does running the program locally allow you to better control the information that it trains on?
in a sense: if you don't let it connect to the internet, it won't be able to take your data to the creators