1537
you are viewing a single comment's thread
view the rest of the comments
[-] brucethemoose@lemmy.world 331 points 1 month ago* (last edited 1 month ago)

As a fervent AI enthusiast, I disagree.

...I'd say it's 97% hype and marketing.

It's crazy how much fud is flying around, and legitimately buries good open research. It's also crazy what these giant corporations are explicitly saying what they're going to do, and that anyone buys it. TSMC's allegedly calling Sam Altman a 'podcast bro' is spot on, and I'd add "manipulative vampire" to that.

Talk to any long-time resident of localllama and similar "local" AI communities who actually dig into this stuff, and you'll find immense skepticism, not the crypto-like AI bros like you find on linkedin, twitter and such and blot everything out.

[-] falkerie71@sh.itjust.works 100 points 1 month ago

For real. Being a software engineer with basic knowledge in ML, I'm just sick of companies from every industry being so desperate to cling onto the hype train they're willing to label anything with AI, even if it has little or nothing to do with it, just to boost their stock value. I would be so uncomfortable being an employee having to do this.

[-] Mikelius@lemmy.world 32 points 1 month ago

For sure, it seems like 90% of ai startups are nothing more than front end wrappers for a gpt instance.

[-] dan@upvote.au 20 points 1 month ago* (last edited 1 month ago)

They're all built on top of OpenAI which is very unprofitable at the moment. Feels like the whole industry is built on a shaky foundation.

Putting the entire fate of your company in a different company (OpenAI) is not a great business move. I guess the successful AI startups will eventually transition to self-hosted models like Llama, if they survive that long.

[-] Zos_Kia@lemmynsfw.com 6 points 1 month ago

Most projects I've been in contact with are very aware of that fact. That's why telemetry is so big right now. Everybody is building datasets in the hopes of fine tuning smaller, cheaper models once they have enough good quality data.

[-] xavier666@lemm.ee 6 points 1 month ago

My company is realizing that hosting a model which will be private, cost-effective, and performing better than traditional algorithms is like finding a unicorn. Few months back, the top execs were jumping around GenAI like a bunch of kids. Fortunately, the Sr. research head beat some sense into them.

[-] falkerie71@sh.itjust.works 2 points 1 month ago

You're lucky there's a higher up that could talk down the even higher ups. Though, sometimes it's not even about the r&d teams.

I saw company wide HR educational emails or courses telling you how to improve you work quality/efficiency, and one of them tells us to "research AI" and learn how to utilize it, talking about how great it is and improved the work efficiency by 30%. Sure, it has its uses, but I won't go touting how great it is. And with how ChatGPT works, you have to be the biggest idiot in the world to upload all your sensitive stuff to ChatGPT just for it to make a spreadsheet faster. But without these disclaimers in the email, I doubt regular clerical staff knows about this, and it's extremely dangerous.

[-] Zos_Kia@lemmynsfw.com 1 points 1 month ago

What kind of use-cases was it, where you didn't find suitable local models to work with ? I've found that general "chatbot" things are hit and miss but more domain-constrained tasks (such as extracting structured entities from unstructured text) are pretty reliable even on smaller models. I'm not counting my chickens yet as my dataset is still somewhat small but preliminary testing has been very promising in that regard.

[-] xavier666@lemm.ee 2 points 1 month ago

What kind of use-cases was it, where you didn’t find suitable local models to work with ?

Any time you ask very domain specific questions; eg "i have collected some soil samples from the mesolithic age near the Amazon basin which have high sulfur and phosphorus content compared to my other samples. What factors could contribute to this distribution?", both of-the-shelf local models & OpenAI fail.

The main reason is because these models are not trained on highly-specialized domains of text. Sometimes the models start hallucinating and which reduces our trust upon them.

[-] Zos_Kia@lemmynsfw.com 2 points 1 month ago

“i have collected some soil samples from the mesolithic age near the Amazon basin which have high sulfur and phosphorus content compared to my other samples. What factors could contribute to this distribution?”

Haha yeah the top execs were tripping balls if they thought some off-the-shelf product would be able to answer this kind of expert questions. That's like trying to replace an expert craftsman with a 3D printer.

[-] Badland9085@lemm.ee 6 points 1 month ago

As someone who was working really hard trying to get my company to be able use some classical ML (with very limited amounts of data), with some knowledge on how AI works, and just generally want to do some cool math stuff at work, being asked incessantly to shove AI into any problem that our execs think are “good sells” and be pressured to think about how we can “use AI” was a terrible feel. They now think my work is insufficient and has been tightening the noose on my team.

[-] falkerie71@sh.itjust.works 3 points 1 month ago

This. Exactly.

load more comments (56 replies)
this post was submitted on 28 Oct 2024
1537 points (98.7% liked)

Technology

60086 readers
2410 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS