Proton's very biased article on Deepseek
-
[email protected]replied to [email protected] last edited by
I certainly think it's cool, but the further you stray from the beaten path, the more newly janky it gets. I'm sure there's a good workflow here, it'll just take some time to find it.
-
[email protected]replied to [email protected] last edited by
They are absolutely right! Most people don't give a fuck about hosting their own AI, they just download "Deepsneak" and chat..and it is unfortunately even worse than "ClosedAI", cuz they are based in China. Thats why I hope Duckduckgo will host deepseek on their servers (as it is very lightweight in resources, yes?), then we will all benefit from it.
-
[email protected]replied to [email protected] last edited by
Ok correction noted, It still makes DeepSeek look better
-
[email protected]replied to [email protected] last edited by
When the CEO praises Trump, says China bad because China while hiding that occidental AIs have the same kind of censorship, that’s hypocrisy.
-
[email protected]replied to [email protected] last edited by
Billionaires are really pissed about it, I’m happy.
-
[email protected]replied to [email protected] last edited by
failing to make it clear that the reason people love Deepseek is that you can download and it run it securely on any of your own private devices or servers
That's not why. Almost no one is going to do that. That's why they didn't mention it.
-
[email protected]replied to [email protected] last edited by
You should try the comparison between the larger models and the distilled models yourself before you make judgment. I suspect you're going to be surprised by the output.
All of the models are basically generating possible outcomes based on noise. So if you ask it the same model the same question five different times and five different sessions you're going to get five different variations on an answer.
You will find that an x out of five score between models is not that significantly different.
For certain cases larger models are advantageous. If you need a model to return a substantial amount of content to you. If you're asking it to write you a chapter story. Larger models will definitely give you better output and better variation.
But if you're asking you to help you with a piece of code or explain some historical event to you, The average 14B model that will fit on any computer with a video card will give you a perfectly serviceable answer.
-
[email protected]replied to [email protected] last edited by
Yet not great from a privacy perspective. They don't even allow third party email apps.
-
[email protected]replied to [email protected] last edited by
The article goes into great detail about how it's different from OpenAI so, no.
-
[email protected]replied to [email protected] last edited by
Serious question, how does them being based in China make them worse? I'd much rather have a foreign intelligence agency collect data on me than one in the country in which I live. It's not like I'd get extradited to China.
-
[email protected]replied to [email protected] last edited by
You don't think ChatGPT reflects western propaganda?
-
[email protected]replied to [email protected] last edited by
Probably, but that's not the new thing
-
[email protected]replied to [email protected] last edited by
He did not praise trump though
-
[email protected]replied to [email protected] last edited by
hiding that occidental AIs have the same kind of censorship
This is the second sentence in the article:
AI chat apps like ChatGPT collect user data, filter responses, and make content moderation decisions that are not always transparent.
They go into great deal about how they actually do not have the same kind of censorship. You should try reading the article before commenting on it.
-
[email protected]replied to [email protected] last edited by
True, hosting deepseek yourself is much better. I'd still wait and see if anyone finds weird stuff in the code itself but tbh idk how long that could take.
Can't wait for the models to get better and hopefully stay open source!
-
[email protected]replied to [email protected] last edited by
DeepSeek is open source, meaning you can modify code[...] on your own app to create an independent — and more secure — version. However, using DeepSeek in its current form — as it exists today, hosted in China — comes with serious risks for anyone concerned about their most sensitive, private information.
They are not wrong here.
After having read the article fully it doesn't seem to be that partial and acknowledge also the failing of others. It is not as stupid as the CEO stance on "Republicans helping the little guys" for sure.
-
[email protected]replied to [email protected] last edited by
Sure it might but the thing is it may still acknowledge that there are different opinions on some topics. Does reflect how whilst governments may have a narrative, people can say what they think. In China, that's a different story...
-
[email protected]replied to [email protected] last edited by
But his 'support' of the republicans was saying that 10 years ago they used to be against big tech and that he hoped Trump would vary that forward. Obviously Trump is very unlikely to do this but he is literally just hoping the republicans would do something about big tech that the dema didn't do
-
[email protected]replied to [email protected] last edited by
They didn't really praise them. They just hoped that the republicans would go back to being against big tech (like they used to be 10 years ago he claims). Obviously, Trump's not going to do that but I think we can all agree big tech is a big problem
-
[email protected]replied to [email protected] last edited by
I have tried them, and to be honest I was not surprised. The hosted service was better at longer code snippets and in particular, I found that it was consistently better at producing valid chain of thought reasoning chains (I've found that a lot of simpler models, including the distills, tend to produce shallow reasoning chains, even when they get the answer to a question right).
I'm aware of how these models work; I work in this field and have been developing a benchmark for reasoning capabilities in LLMs. The distills are certainly still technically impressive and it's nice that they exist, but the gap between them and the hosted version is unfortunately nontrivial.