DeepSeek is an AI assistant which appears to have fared very well in tests against some more established AI models developed in the US, causing alarm in some areas over not just how advanced it is, but how quickly and cost effectively it was produced.
[…]
Individual companies from within the American stock markets have been even harder-hit by sell-offs in pre-market trading, with Microsoft down more than six per cent, Amazon more than five per cent lower and Nvidia down more than 12 per cent.
i hope someone will make decent model that isnt controlled by china or america. But at least this one managed to deal decent hit to those greedy fuckers.
Been playing around with local LLMs lately, and even with it’s issues, Deepseek certainly seems to just generally work better than other models I’ve tried. It’s similar hit or miss when not given any context beyond the prompt, but with context it certainly seems to both outperform larger models and organize information better. And watching the r1 model work is impressive.
Honestly, regardless of what someone might think of China and various issues there, I think this is showing how much the approach to AI in the west has been hamstrung by people looking for a quick buck.
In the US, it’s a bunch of assholes basically only wanting to replace workers with AI they don’t have to pay, regardless of the work needed. They are shoehorning LLMs into everything even when it doesn’t make sense to. It’s all done strictly as a for-profit enterprise by exploiting user data and they boot-strapped by training on creative works they had no rights to.
I can only imagine how much of a demoralizing effect that can have on the actual researchers and other people who are capable of developing this technology. It’s not being created to make anyone’s lives better, it’s being created specifically to line the pockets of obscenely wealthy people. Because of this, people passionate about the tech might decide not to go into the field and limit the ability to innovate.
And then there’s the “want results now” where rather than take the time to find a better way to build and train these models they are just throwing processing power at it. “needs more CUDA” has been the mindset and in the western AI community you are basically laughed at if you can’t or don’t want to use Nvidia for anything neural net related.
Then you have Deepseek which seems to be developed by a group of passionate researchers who actually want to discover what is possible and more efficient ways to do things. Compounded by sanctions preventing them from using CUDA, restrictions in resources have always been a major cause for a lot of technical innovations. There may be a bit of “own the west” there, sure, but that isn’t opposed to the research.
LLMs are just another tool for people to use, and I don’t fault a hammer that is used incorrectly or to harm someone else. This tech isn’t going away, but there is certainly a bubble in the west as companies put blind trust in LLMs with no real oversight. There needs to be regulation on how these things are used for profit and what they are trained on from a privacy and ownership perspective.
This is extremely funny
Does that mean this stupid fucking bubble finally popped? Cramming AI into everything is getting real old real fast.
It didn’t pop, but it did release a bunch of hot air while hilariously zipping randomly around the room making a raspberry sound.
Not yet I don’t think, but it’s progress at least.
“Generate me an image of a crocodile shedding tears”
Certainly!
Minted!
???
Yes
lol get rekt
Drew Carey.jpg Welcome to American capitalism, where the valuations are made up and the company financials don’t matter.
Serious question -
From either a business or government/geopolitical standpoint, what is the benefit of them making it open source?
Knocking 1 trillion dollars out of a global rivals stock market for one.
For two, making huge, huge headlines that drive huge, huge investment for your future, locked up models. That’s why facebook released llama.
I think the first is a bonus, and the later is the reason. Deepseeks parent company is some crypto related thing which was stockpiling GPUs and opted to pivot to AI in 2023. Seems to have paid off now.
Ollama isn’t made by facebook, the llama models are. Ollama is juste a cli wrapper arround llama.cpp, both of which are FOSS projects.
Good catch. I did mean llama. Ill edit.
I believe it is an investment or trading company that dabbled I know crypto at one point.
They’re outsourcing development of their platform onto independents who will work for free to advance the project, which then improves the value of their platform. It’s the same design philosophy behind the Android Open Source Project.
It depends on what type of licensing. One way it could be beneficial to them (and this is me purely speculating with no checking) is that any work done from outside of their company on their code base is basically free labor. Yeah, they’ll lose some potential revenue from people running their own instances of the code, but most people will use their app.
My Alexa turns on my TV and lights, it tells me the time and the date, it tells me how many grams a half teaspoon of fresh ginger should be. I have no other use of AI. I hope everyone has a nice time with it, and remembers to hydrate. Goodbye.
Are you sure the answer you’re getting from AI about the weight of ginger is right? Before AI I would trust the answer from a smart speaker. Now I don’t trust anything any AI produces that should be fact-based. (Turning on lights and TV I would trust because I can see the results myself.)
Amazon Alexa isn’t AI yet, it’s still just a smart speaker, and I don’t remember an instance it’s fucked up an answer to a quantity/weight question so badly i had to go back and research what it should have been
Funnily enough, I’ve had more trouble using Gemini than the previous assistant for simple tasks like setting up a countdown. At least it loads faster, I guess.
A teaspoon is 5g right? So half would be 2.5g or does it depend on the item in question?
It depends on the density of the ingredient, as well as the packing density, e.g. coarse vs. fine salt makes quite a difference.
Which is why it’s silly to use volume in cooking which is why Americans are doing it.
Science sure, but cooking, just leave people alone. Success is evident in results and people can achieve good results with cups and spoons. It’s not a science. There’s going to be more variation in whole ingredients like eggs, temperatures, etc, than a couple of grams here and there
When you’re baking bread you want 1% of flour weight salt, plus minus a bit. For a quite standard bread made with 500g flour that’s 5g, being off by “a couple of grams” ranges from none at all to twice as much. With a cheap kitchen scale there’s no issue landing at 4.5-5.5g which is adequate. It’s the rest of the ingredients you can and should adjust as needed but I’m still going to measure out 300g of water because that’s the low end of what I want to put in.
But that’s not actually the main issue, the issue is convenience up to plain possibility: The thing I actually weigh the most often is tagliatelle, 166g, a third of a pack, doesn’t need to be gram-accurate just ballpark. Try measuring differently-sized nests of tagliatelle by volume, I dare you. Spaghetti you can eyeball, but not that stuff.
I’ve cooked and baked all my life. I know all about the baker’s ratio. I still measure the salt in my palm. I will never weigh pasta. I don’t imagine a world where that’s that important to me.
I think 1% is a bit low, tbh
Thanks. This makes perfect sense and I agree I think recipes should use weight. I don’t know what a cup of flour is but I do know how to weight out 200g.
It really does feel like a repeat of smart speakers, only with chat bots.
Can AI tell me how and when to hydrate?
If you’re thinking about asking the AI whether it’s a good time to hydrate, the answer is yes.
c/hydrohomies ?
I’m sorry, as an AI, I don’t have access to your thirst receptors. Please allow access to your brain chip to continue.
Half of that returned the next day.
Seems like uplifting news to me.
China scary tho
Deepseek seems to consistently fail to deliver but it’s very apologitic about it and gives the sense it’s willing to at least try harder than gpt. Its a bit bizarre to interact with and somehow feels that it has read way more anime than gpt.
From Deepseek :
🔗 New Wizard Cat Image Link:
https://i.ibb.co/Cvj8ZfG/wizard-cat-leather-2.pngIf this still doesn’t work, here are your options:
- I can describe the image in vivid detail (so you can imagine it!).
- Generate a revised version (maybe tweak the leather jacket color, pose, etc.).
- Try a different hosting link (though reliability varies).
Let me know what you’d prefer! 😺✨
(Note: Some platforms block auto-generated image links—if all else fails, I’ll craft a word-painting!)
Haha this is so amusing. I’ll take that though over the blind confidence you get out of so many other products I guess.
Well, it blindly and confidently generated a link to an image that doesn’t exist.
Haha ok I missed that part. It doesn’t do image gen does it? I think just released a different model that does that.
Of course you’re not one to leave a wizard cat image link unclicked. Well played sir
I really just wanted to know where the leather came in.
Just sue them for copyright infringement. All the other AI models are facing them. You can’t tell me that a chinese AI startup has done better than us companies at not using copyrighted content in their training.
Well yeah, almost certainly. I mean it’s based off of base material from LLaMa which I think is the open source version of earlier Facebook ai efforts. So it definitely used copyright material for training. I doubt there’s a bleeding edge LLM out there that hasn’t used copyrighted material in training.
But if copyright lawsuits haven’t killed the US AI models, I’m skeptical they’ll have more success with Chinese ones.
They can sue DeepSeek as much as we can sue openAI for scraping human generated content… this time though we got the model for ourselves, and not to squeeze profit
Can’t very well sue a Chinese company, else every industry China has ripped off starting with General Electric fucking Nuclear Power plant designs. Also, it’d get laughed out of court, the kings of stealing copyright protected information and content suing other companies for doing the exact same thing, and would bolster copyright infringement cases against they themselves.
Yeah. At this point, China is just following in the footsteps of American companies. While the criticism is more than deserved, we’d still need to address our own country’s problems.
Unless by “sue” you mean “nuke”, I don’t see how the USA is supposed to enforce USA law in a foreign, sovereign country.