22
7
[-] Lugh@futurology.today 3 points 1 day ago

But a domain expert like a doctor or an accountant is way much more accurate

Actually, not so.

If the AI is trained on narrow data sets, then it beats humans. There's quite a few examples of this recently with different types of medical expertise.

[-] Lugh@futurology.today 8 points 1 day ago

Australia is a paradox in the renewables transition. It is already at 35% for renewable electricity, and has targeted 82% for 2030. Yet it's still a major exporter of coal. Australia exported $127.4 billion worth of coal in 2022-23, and its economy is highly dependent on mining of all types.

It doesn't have much homegrown manufacturing and is committed to eliminating tariffs on Chinese imports. This means of Western countries it might be among the quickest to abandon ICE cars, as it will have access to all the super-cheap Chinese EV's. Especially as it's rolling out infrastructure like this.

23
[-] Lugh@futurology.today 4 points 1 day ago* (last edited 1 day ago)

I still see even the more advanced AIs make simple errors on facts all the time....

47
[-] Lugh@futurology.today 2 points 3 days ago

True. It's what keeps me optimistic. If we can get through the decay of the old world intact, there's a world of post-scarcity plenty ahead.

[-] Lugh@futurology.today 5 points 3 days ago

I find they are good for creative tasks. Picture and music generation, but also ideas - say, give me 10 possible character names for a devious butler in a 1930s murder mystery novel.

But yes, terrible for facts, even rudimentary ones. I get so many errors with this approach its effectively useless.

However, I can see on narrower training data, say genetics, this might be less of a problem.

48
84
21
[-] Lugh@futurology.today 1 points 4 days ago

There's a few ways they say it may help, this one seems the main one.

We foresee a future in which LLMs serve as forward-looking generative models of the scientific literature. LLMs can be part of larger systems that assist researchers in determining the best experiment to conduct next. One key step towards achieving this vision is demonstrating that LLMs can identify likely results. For this reason, BrainBench involved a binary choice between two possible results. LLMs excelled at this task, which brings us closer to systems that are practically useful. In the future, rather than simply selecting the most likely result for a study, LLMs can generate a set of possible results and judge how likely each is. Scientists may interactively use these future systems to guide the design of their experiments.

[-] Lugh@futurology.today 2 points 4 days ago* (last edited 4 days ago)

It’s only open source if the training data is and it probably isn’t, is it?

I don't know, though DeepSeek talk of theirs being "fully" open-source.

Part of the advantage of doing this (apart from helping bleed your rivals dry) is to get the benefit of others working on your model. So it makes sense to maximise openness and access.

22
24

Alibaba's Qwen team just released QwQ-32B-Preview, a powerful new open-source AI reasoning model that can reason step-by-step through challenging problems and directly competes with OpenAI's o1 series across benchmarks.

The details:

QwQ features a 32K context window, outperforming o1-mini and competing with o1-preview on key math and reasoning benchmarks.

The model was tested across several of the most challenging math and programming benchmarks, showing major advances in deep reasoning.

QwQ demonstrates ‘deep introspection,’ talking through problems step-by-step and questioning and examining its own answers to reason to a solution.

The Qwen team noted several issues in the Preview model, including getting stuck in reasoning loops, struggling with common sense, and language mixing.

Why it matters: Between QwQ and DeepSeek, open-source reasoning models are here — and Chinese firms are absolutely cooking with new models that nearly match the current top closed leaders. Has OpenAI’s moat dried up, or does the AI leader have something special up its sleeve before the end of the year?

14

By fine-tuning models with carefully curated techniques, TTT improved accuracy sixfold in some cases and set a new state-of-the-art for purely neural approaches, achieving 53% accuracy with an 8-billion-parameter model. When combined with program synthesis methods, the models reached 61.9% accuracy, matching average human performance. The findings suggest that symbolic reasoning isn’t essential for solving complex problems, emphasizing the power of dynamic, computation-focused approaches during inference. → Read the full paper here.

[-] Lugh@futurology.today 1 points 5 days ago* (last edited 5 days ago)

I lived in Hong Kong for a few years. It has superlative public transport, and the (human) taxis were reasonably priced. However, as its so densely populated, I can only see cars getting so much traction. After a certain point the traffic jams are unavoidable.

7
34

This article - How the Rise of New Digital Workers Will Lead to an Unlimited Age - makes the mainstream case for the future of employment with respect to robotics and AI. By mainstream, I mean that it completely ignores the central question. What happens to human employees when most or all (even future uninvented) work can be done for pennies an hour by AI & robotics employees?

As almost always, he poses the question, and in classic Strawman fashion - pretends to answer it, by answering a different question. Mr Benioff says automation has always created more jobs than it eliminates. But that only answers a different question and ignores the most important one.

Mr. Benioff, CEO of Salesforce and owner of TIME magazine is no different from mainstream economists, or the Silicon Valley elite, in building this world and blindly leading us to it.

One day society is going to have to wake up to the fact we are being duped by these people, and the longer we keep believing them, the more we just get all the angst and chaos, and none of the understanding we need to fashion a new reality.

[-] Lugh@futurology.today 3 points 6 days ago

I don't know the specifics. What seems more relevant to me is that lots of automakers around the world are getting to Level 4 by various, mostly similar ways.

Once you have Level 4 you have a viable robotaxi business model. Even if you stick to geo-fenced areas, and mapped routes, that covers 80%+ of urban taxi journeys.

The same holds true for buses and public transit. I'm very interested to see how efforts like this Level 4 mini-shuttle bus in France progress.

https://www.euronews.com/next/2024/10/26/this-self-driving-shuttle-transports-people-in-rural-france-is-it-the-future-of-mobility

When robotaxis & mass transit like these are common, how many people will still want private cars?

[-] Lugh@futurology.today 189 points 8 months ago

Good news for pigs. I'll be delighted to see factory farming disappear and be replaced by tech like this.

[-] Lugh@futurology.today 196 points 1 year ago* (last edited 1 year ago)

I think fediverse people are wildly overestimating how much 99% of Reddit users care about this. The mod team on r/futurology (I'm one of them) set up a fediverse site just over a month ago (here you go - https://futurology.today/ ) It's been modestly successful so far, but the vast majority of subscribers seem to be coming from elsewhere in the fediverse, not migrants from Reddit.

This is despite the fact we've permanently stickied a post to the top of the sub. r/futurology has over 19 million subscribers, and yet the fediverse is only attracting a tiny trickle of them. I doubt most people on Reddit even know what the word fediverse means.

view more: next ›

Lugh

joined 1 year ago
MODERATOR OF