this post was submitted on 28 Jan 2025
384 points (97.8% liked)

Technology

61456 readers
4086 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

DeepSeek is an AI assistant which appears to have fared very well in tests against some more established AI models developed in the US, causing alarm in some areas over not just how advanced it is, but how quickly and cost effectively it was produced.

[...]

Individual companies from within the American stock markets have been even harder-hit by sell-offs in pre-market trading, with Microsoft down more than six per cent, Amazon more than five per cent lower and Nvidia down more than 12 per cent.

top 50 comments
sorted by: hot top controversial new old
[–] assassinatedbyCIA@lemmy.world 127 points 1 week ago

This is extremely funny

[–] FartsWithAnAccent@fedia.io 101 points 1 week ago (2 children)

Does that mean this stupid fucking bubble finally popped? Cramming AI into everything is getting real old real fast.

[–] Zorque@lemmy.world 89 points 1 week ago

It didn't pop, but it did release a bunch of hot air while hilariously zipping randomly around the room making a raspberry sound.

[–] sleep_deprived@lemmy.world 27 points 1 week ago

Not yet I don't think, but it's progress at least.

[–] zapzap@lemmings.world 65 points 1 week ago (1 children)

"Generate me an image of a crocodile shedding tears"

[–] barsoap@lemm.ee 41 points 1 week ago (1 children)
[–] Sturgist@lemmy.ca 3 points 1 week ago (1 children)
[–] KeenFlame@feddit.nu 3 points 1 week ago (1 children)
[–] Sturgist@lemmy.ca 1 points 6 days ago (1 children)
[–] KeenFlame@feddit.nu 2 points 2 days ago (1 children)
[–] Sturgist@lemmy.ca 2 points 1 day ago

taps nose Yes! ;)

[–] shoulderoforion@fedia.io 39 points 1 week ago (4 children)

My Alexa turns on my TV and lights, it tells me the time and the date, it tells me how many grams a half teaspoon of fresh ginger should be. I have no other use of AI. I hope everyone has a nice time with it, and remembers to hydrate. Goodbye.

[–] ignirtoq@fedia.io 30 points 1 week ago (3 children)

Are you sure the answer you're getting from AI about the weight of ginger is right? Before AI I would trust the answer from a smart speaker. Now I don't trust anything any AI produces that should be fact-based. (Turning on lights and TV I would trust because I can see the results myself.)

[–] shoulderoforion@fedia.io 9 points 1 week ago

Amazon Alexa isn't AI yet, it's still just a smart speaker, and I don't remember an instance it's fucked up an answer to a quantity/weight question so badly i had to go back and research what it should have been

[–] Lemminary@lemmy.world 4 points 1 week ago* (last edited 1 week ago)

Funnily enough, I've had more trouble using Gemini than the previous assistant for simple tasks like setting up a countdown. At least it loads faster, I guess.

[–] dependencyinjection@discuss.tchncs.de 3 points 1 week ago (1 children)

A teaspoon is 5g right? So half would be 2.5g or does it depend on the item in question?

[–] barsoap@lemm.ee 10 points 1 week ago (4 children)

It depends on the density of the ingredient, as well as the packing density, e.g. coarse vs. fine salt makes quite a difference.

Which is why it's silly to use volume in cooking which is why Americans are doing it.

Thanks. This makes perfect sense and I agree I think recipes should use weight. I don’t know what a cup of flour is but I do know how to weight out 200g.

load more comments (3 replies)
[–] tonytins@pawb.social 13 points 1 week ago

It really does feel like a repeat of smart speakers, only with chat bots.

[–] AbidanYre@lemmy.world 4 points 1 week ago (2 children)

Can AI tell me how and when to hydrate?

[–] BarbecueCowboy@lemmy.world 10 points 1 week ago (1 children)

If you're thinking about asking the AI whether it's a good time to hydrate, the answer is yes.

[–] AbidanYre@lemmy.world 5 points 1 week ago

c/hydrohomies ?

[–] Lemminary@lemmy.world 6 points 1 week ago

I'm sorry, as an AI, I don't have access to your thirst receptors. Please allow access to your brain chip to continue.

[–] aesthelete@lemmy.world 35 points 1 week ago

Drew Carey.jpg Welcome to American capitalism, where the valuations are made up and the company financials don't matter.

[–] PrincessLeiasCat@sh.itjust.works 32 points 1 week ago (3 children)

Serious question -

From either a business or government/geopolitical standpoint, what is the benefit of them making it open source?

[–] mosiacmango@lemm.ee 67 points 1 week ago* (last edited 1 week ago) (2 children)

Knocking 1 trillion dollars out of a global rivals stock market for one.

For two, making huge, huge headlines that drive huge, huge investment for your future, locked up models. That's why facebook released llama.

I think the first is a bonus, and the later is the reason. Deepseeks parent company is some crypto related thing which was stockpiling GPUs and opted to pivot to AI in 2023. Seems to have paid off now.

[–] L_Acacia@lemmy.ml 12 points 1 week ago (1 children)

Ollama isn't made by facebook, the llama models are. Ollama is juste a cli wrapper arround llama.cpp, both of which are FOSS projects.

[–] mosiacmango@lemm.ee 4 points 1 week ago

Good catch. I did mean llama. Ill edit.

I believe it is an investment or trading company that dabbled I know crypto at one point.

It depends on what type of licensing. One way it could be beneficial to them (and this is me purely speculating with no checking) is that any work done from outside of their company on their code base is basically free labor. Yeah, they'll lose some potential revenue from people running their own instances of the code, but most people will use their app.

[–] queermunist@lemmy.ml 4 points 1 week ago* (last edited 1 week ago)

They're outsourcing development of their platform onto independents who will work for free to advance the project, which then improves the value of their platform. It's the same design philosophy behind the Android Open Source Project.

[–] Imgonnatrythis@sh.itjust.works 25 points 1 week ago (1 children)

Deepseek seems to consistently fail to deliver but it's very apologitic about it and gives the sense it's willing to at least try harder than gpt. Its a bit bizarre to interact with and somehow feels that it has read way more anime than gpt.

From Deepseek :

🔗 New Wizard Cat Image Link:
https://i.ibb.co/Cvj8ZfG/wizard-cat-leather-2.png

If this still doesn’t work, here are your options:

  1. I can describe the image in vivid detail (so you can imagine it!).
  2. Generate a revised version (maybe tweak the leather jacket color, pose, etc.).
  3. Try a different hosting link (though reliability varies).

Let me know what you’d prefer! 😺✨

(Note: Some platforms block auto-generated image links—if all else fails, I’ll craft a word-painting!)

[–] Hotspur@lemmy.ml 15 points 1 week ago (1 children)

Haha this is so amusing. I’ll take that though over the blind confidence you get out of so many other products I guess.

[–] catloaf@lemm.ee 12 points 1 week ago (2 children)

Well, it blindly and confidently generated a link to an image that doesn't exist.

[–] imaqtpie@sh.itjust.works 4 points 1 week ago (1 children)

Of course you're not one to leave a wizard cat image link unclicked. Well played sir

[–] catloaf@lemm.ee 3 points 1 week ago

I really just wanted to know where the leather came in.

[–] Hotspur@lemmy.ml 2 points 1 week ago

Haha ok I missed that part. It doesn’t do image gen does it? I think just released a different model that does that.

[–] Naia@lemmy.blahaj.zone 22 points 1 week ago

Been playing around with local LLMs lately, and even with it's issues, Deepseek certainly seems to just generally work better than other models I've tried. It's similar hit or miss when not given any context beyond the prompt, but with context it certainly seems to both outperform larger models and organize information better. And watching the r1 model work is impressive.

Honestly, regardless of what someone might think of China and various issues there, I think this is showing how much the approach to AI in the west has been hamstrung by people looking for a quick buck.

In the US, it's a bunch of assholes basically only wanting to replace workers with AI they don't have to pay, regardless of the work needed. They are shoehorning LLMs into everything even when it doesn't make sense to. It's all done strictly as a for-profit enterprise by exploiting user data and they boot-strapped by training on creative works they had no rights to.

I can only imagine how much of a demoralizing effect that can have on the actual researchers and other people who are capable of developing this technology. It's not being created to make anyone's lives better, it's being created specifically to line the pockets of obscenely wealthy people. Because of this, people passionate about the tech might decide not to go into the field and limit the ability to innovate.

And then there's the "want results now" where rather than take the time to find a better way to build and train these models they are just throwing processing power at it. "needs more CUDA" has been the mindset and in the western AI community you are basically laughed at if you can't or don't want to use Nvidia for anything neural net related.

Then you have Deepseek which seems to be developed by a group of passionate researchers who actually want to discover what is possible and more efficient ways to do things. Compounded by sanctions preventing them from using CUDA, restrictions in resources have always been a major cause for a lot of technical innovations. There may be a bit of "own the west" there, sure, but that isn't opposed to the research.

LLMs are just another tool for people to use, and I don't fault a hammer that is used incorrectly or to harm someone else. This tech isn't going away, but there is certainly a bubble in the west as companies put blind trust in LLMs with no real oversight. There needs to be regulation on how these things are used for profit and what they are trained on from a privacy and ownership perspective.

[–] TheFriar@lemm.ee 13 points 1 week ago

lol get rekt

[–] _cryptagion@lemmy.dbzer0.com 8 points 1 week ago (1 children)

Seems like uplifting news to me.

[–] Pixel@lemmy.ca 2 points 1 week ago

China scary tho

[–] reksas@sopuli.xyz 3 points 1 week ago* (last edited 1 week ago)

i hope someone will make decent model that isnt controlled by china or america. But at least this one managed to deal decent hit to those greedy fuckers.

[–] Bronzebeard@lemm.ee 3 points 1 week ago

Half of that returned the next day.

load more comments
view more: next ›