this post was submitted on 28 Jan 2025
380 points (98.0% liked)

Technology

61227 readers
4082 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

DeepSeek is an AI assistant which appears to have fared very well in tests against some more established AI models developed in the US, causing alarm in some areas over not just how advanced it is, but how quickly and cost effectively it was produced.

[...]

Individual companies from within the American stock markets have been even harder-hit by sell-offs in pre-market trading, with Microsoft down more than six per cent, Amazon more than five per cent lower and Nvidia down more than 12 per cent.

top 50 comments
sorted by: hot top controversial new old
[–] assassinatedbyCIA@lemmy.world 126 points 3 days ago

This is extremely funny

[–] FartsWithAnAccent@fedia.io 98 points 3 days ago (2 children)

Does that mean this stupid fucking bubble finally popped? Cramming AI into everything is getting real old real fast.

[–] Zorque@lemmy.world 88 points 3 days ago

It didn't pop, but it did release a bunch of hot air while hilariously zipping randomly around the room making a raspberry sound.

[–] sleep_deprived@lemmy.world 27 points 3 days ago

Not yet I don't think, but it's progress at least.

[–] zapzap@lemmings.world 64 points 3 days ago (1 children)

"Generate me an image of a crocodile shedding tears"

[–] barsoap@lemm.ee 40 points 3 days ago (1 children)
[–] Naia@lemmy.blahaj.zone 21 points 3 days ago

Been playing around with local LLMs lately, and even with it's issues, Deepseek certainly seems to just generally work better than other models I've tried. It's similar hit or miss when not given any context beyond the prompt, but with context it certainly seems to both outperform larger models and organize information better. And watching the r1 model work is impressive.

Honestly, regardless of what someone might think of China and various issues there, I think this is showing how much the approach to AI in the west has been hamstrung by people looking for a quick buck.

In the US, it's a bunch of assholes basically only wanting to replace workers with AI they don't have to pay, regardless of the work needed. They are shoehorning LLMs into everything even when it doesn't make sense to. It's all done strictly as a for-profit enterprise by exploiting user data and they boot-strapped by training on creative works they had no rights to.

I can only imagine how much of a demoralizing effect that can have on the actual researchers and other people who are capable of developing this technology. It's not being created to make anyone's lives better, it's being created specifically to line the pockets of obscenely wealthy people. Because of this, people passionate about the tech might decide not to go into the field and limit the ability to innovate.

And then there's the "want results now" where rather than take the time to find a better way to build and train these models they are just throwing processing power at it. "needs more CUDA" has been the mindset and in the western AI community you are basically laughed at if you can't or don't want to use Nvidia for anything neural net related.

Then you have Deepseek which seems to be developed by a group of passionate researchers who actually want to discover what is possible and more efficient ways to do things. Compounded by sanctions preventing them from using CUDA, restrictions in resources have always been a major cause for a lot of technical innovations. There may be a bit of "own the west" there, sure, but that isn't opposed to the research.

LLMs are just another tool for people to use, and I don't fault a hammer that is used incorrectly or to harm someone else. This tech isn't going away, but there is certainly a bubble in the west as companies put blind trust in LLMs with no real oversight. There needs to be regulation on how these things are used for profit and what they are trained on from a privacy and ownership perspective.

[–] reksas@sopuli.xyz 3 points 2 days ago* (last edited 2 days ago)

i hope someone will make decent model that isnt controlled by china or america. But at least this one managed to deal decent hit to those greedy fuckers.

[–] aesthelete@lemmy.world 35 points 3 days ago

Drew Carey.jpg Welcome to American capitalism, where the valuations are made up and the company financials don't matter.

[–] TheFriar@lemm.ee 13 points 3 days ago

lol get rekt

[–] shoulderoforion@fedia.io 39 points 3 days ago (4 children)

My Alexa turns on my TV and lights, it tells me the time and the date, it tells me how many grams a half teaspoon of fresh ginger should be. I have no other use of AI. I hope everyone has a nice time with it, and remembers to hydrate. Goodbye.

[–] ignirtoq@fedia.io 30 points 3 days ago (3 children)

Are you sure the answer you're getting from AI about the weight of ginger is right? Before AI I would trust the answer from a smart speaker. Now I don't trust anything any AI produces that should be fact-based. (Turning on lights and TV I would trust because I can see the results myself.)

[–] shoulderoforion@fedia.io 9 points 3 days ago

Amazon Alexa isn't AI yet, it's still just a smart speaker, and I don't remember an instance it's fucked up an answer to a quantity/weight question so badly i had to go back and research what it should have been

[–] Lemminary@lemmy.world 4 points 3 days ago* (last edited 3 days ago)

Funnily enough, I've had more trouble using Gemini than the previous assistant for simple tasks like setting up a countdown. At least it loads faster, I guess.

[–] dependencyinjection@discuss.tchncs.de 3 points 3 days ago (1 children)

A teaspoon is 5g right? So half would be 2.5g or does it depend on the item in question?

[–] barsoap@lemm.ee 10 points 3 days ago (2 children)

It depends on the density of the ingredient, as well as the packing density, e.g. coarse vs. fine salt makes quite a difference.

Which is why it's silly to use volume in cooking which is why Americans are doing it.

[–] raef@lemmy.world 1 points 2 days ago* (last edited 2 days ago) (1 children)

Science sure, but cooking, just leave people alone. Success is evident in results and people can achieve good results with cups and spoons. It's not a science. There's going to be more variation in whole ingredients like eggs, temperatures, etc, than a couple of grams here and there

[–] barsoap@lemm.ee 2 points 2 days ago* (last edited 2 days ago) (1 children)

When you're baking bread you want 1% of flour weight salt, plus minus a bit. For a quite standard bread made with 500g flour that's 5g, being off by "a couple of grams" ranges from none at all to twice as much. With a cheap kitchen scale there's no issue landing at 4.5-5.5g which is adequate. It's the rest of the ingredients you can and should adjust as needed but I'm still going to measure out 300g of water because that's the low end of what I want to put in.

But that's not actually the main issue, the issue is convenience up to plain possibility: The thing I actually weigh the most often is tagliatelle, 166g, a third of a pack, doesn't need to be gram-accurate just ballpark. Try measuring differently-sized nests of tagliatelle by volume, I dare you. Spaghetti you can eyeball, but not that stuff.

[–] raef@lemmy.world 1 points 2 days ago* (last edited 2 days ago)

I've cooked and baked all my life. I know all about the baker's ratio. I still measure the salt in my palm. I will never weigh pasta. I don't imagine a world where that's that important to me.

I think 1% is a bit low, tbh

Thanks. This makes perfect sense and I agree I think recipes should use weight. I don’t know what a cup of flour is but I do know how to weight out 200g.

[–] tonytins@pawb.social 13 points 3 days ago

It really does feel like a repeat of smart speakers, only with chat bots.

[–] AbidanYre@lemmy.world 4 points 3 days ago (2 children)

Can AI tell me how and when to hydrate?

[–] BarbecueCowboy@lemmy.world 10 points 3 days ago (1 children)

If you're thinking about asking the AI whether it's a good time to hydrate, the answer is yes.

[–] AbidanYre@lemmy.world 5 points 3 days ago

c/hydrohomies ?

[–] Lemminary@lemmy.world 6 points 3 days ago

I'm sorry, as an AI, I don't have access to your thirst receptors. Please allow access to your brain chip to continue.

[–] PrincessLeiasCat@sh.itjust.works 32 points 3 days ago (3 children)

Serious question -

From either a business or government/geopolitical standpoint, what is the benefit of them making it open source?

[–] mosiacmango@lemm.ee 66 points 3 days ago* (last edited 3 days ago) (2 children)

Knocking 1 trillion dollars out of a global rivals stock market for one.

For two, making huge, huge headlines that drive huge, huge investment for your future, locked up models. That's why facebook released llama.

I think the first is a bonus, and the later is the reason. Deepseeks parent company is some crypto related thing which was stockpiling GPUs and opted to pivot to AI in 2023. Seems to have paid off now.

[–] L_Acacia@lemmy.ml 12 points 3 days ago (1 children)

Ollama isn't made by facebook, the llama models are. Ollama is juste a cli wrapper arround llama.cpp, both of which are FOSS projects.

[–] mosiacmango@lemm.ee 4 points 3 days ago

Good catch. I did mean llama. Ill edit.

I believe it is an investment or trading company that dabbled I know crypto at one point.

[–] queermunist@lemmy.ml 6 points 3 days ago* (last edited 3 days ago)

They're outsourcing development of their platform onto independents who will work for free to advance the project, which then improves the value of their platform. It's the same design philosophy behind the Android Open Source Project.

It depends on what type of licensing. One way it could be beneficial to them (and this is me purely speculating with no checking) is that any work done from outside of their company on their code base is basically free labor. Yeah, they'll lose some potential revenue from people running their own instances of the code, but most people will use their app.

[–] Imgonnatrythis@sh.itjust.works 25 points 3 days ago (1 children)

Deepseek seems to consistently fail to deliver but it's very apologitic about it and gives the sense it's willing to at least try harder than gpt. Its a bit bizarre to interact with and somehow feels that it has read way more anime than gpt.

From Deepseek :

🔗 New Wizard Cat Image Link:
https://i.ibb.co/Cvj8ZfG/wizard-cat-leather-2.png

If this still doesn’t work, here are your options:

  1. I can describe the image in vivid detail (so you can imagine it!).
  2. Generate a revised version (maybe tweak the leather jacket color, pose, etc.).
  3. Try a different hosting link (though reliability varies).

Let me know what you’d prefer! 😺✨

(Note: Some platforms block auto-generated image links—if all else fails, I’ll craft a word-painting!)

[–] Hotspur@lemmy.ml 15 points 3 days ago (1 children)

Haha this is so amusing. I’ll take that though over the blind confidence you get out of so many other products I guess.

[–] catloaf@lemm.ee 12 points 3 days ago (2 children)

Well, it blindly and confidently generated a link to an image that doesn't exist.

[–] imaqtpie@sh.itjust.works 4 points 3 days ago (1 children)

Of course you're not one to leave a wizard cat image link unclicked. Well played sir

[–] catloaf@lemm.ee 3 points 3 days ago

I really just wanted to know where the leather came in.

[–] Hotspur@lemmy.ml 2 points 3 days ago

Haha ok I missed that part. It doesn’t do image gen does it? I think just released a different model that does that.

[–] _cryptagion@lemmy.dbzer0.com 8 points 3 days ago (1 children)

Seems like uplifting news to me.

[–] Pixel@lemmy.ca 2 points 2 days ago

China scary tho

[–] Bronzebeard@lemm.ee 3 points 2 days ago

Half of that returned the next day.

load more comments
view more: next ›