jj4211

joined 2 years ago
[–] jj4211@lemmy.world 6 points 1 day ago (1 children)

Bad for business, but perhaps not as bad for the rich business owners that can capitalize on the market chaos a bit without even pretending to contribute to actual productivity.

[–] jj4211@lemmy.world 12 points 1 day ago

Yeah it's essentially letting China win by default..

[–] jj4211@lemmy.world 5 points 1 day ago (1 children)

That's one pair of philosophies that creep me out both ways. Both the anti natalists and pro natalists.

Deciding for yourself is one thing, imposing your choice on others is maddening.

I don't know if the comment quite raises to the level of anti natalist though. Maybe it's grading on a curve of reading some more hard core anti natalists, but that comment felt tame and felt like they wouldn't necessarily object to a couple having one child or even two, being somewhat below the replacement level..

[–] jj4211@lemmy.world 2 points 1 day ago

Thanks for that write up, very informative about what went down.

I wonder about having different alert sounds. The one alert sound I barely think to take seriously. I read them and I think I would notice unique phrasing, but I also imagine people are tempted by the ability to turn off emergency alerts as they seem a bit overused.

[–] jj4211@lemmy.world 18 points 2 days ago

The issue here is that we've well gone into sharply exponential expenditure of resources for reduced gains and a lot of good theory predicting that the breakthroughs we have seen are about tapped out, and no good way to anticipate when a further breakthrough might happen, could be real soon or another few decades off.

I anticipate a pull back of resources invested and a settling for some middle ground where it is absolutely useful/good enough to have the current state of the art, mostly wrong but very quick when it's right with relatively acceptable consequences for the mistakes. Perhaps society getting used to the sorts of things it will fail at and reducing how much time we try to make the LLMs play in that 70% wrong sort of use case.

I see LLMs as replacing first line support, maybe escalating to a human when actual stakes arise for a call (issuing warranty replacement, usage scenario that actually has serious consequences, customer demanding the human escalation after recognizing they are falling through the AI cracks without the AI figuring out to escalate). I expect to rarely ever see "stock photography" used again. I expect animation to employ AI at least for backgrounds like "generic forest that no one is going to actively look like, but it must be plausibly forest". I expect it to augment software developers, but not able to enable a generic manager to code up whatever he might imagine. The commonality in all these is that they live in the mind numbing sorts of things current LLM can get right and/or a high tolerance for mistakes with ample opportunity for humans to intervene before the mistakes inflict much cost.

[–] jj4211@lemmy.world 2 points 2 days ago

Well, here's me pinning my hopes on your interpretation. A few more moderate leaders in the world would be a gigantic relief after so many years of how things have been going. I mostly grew up the last time the world swung a bit more moderate and would be ecstatic to feel that way again.

[–] jj4211@lemmy.world 4 points 2 days ago

I've found that as an ambient code completion facility it's... interesting, but I don't know if it's useful or not...

So on average, it's totally wrong about 80% of the time, 19% of the time the first line or two is useful (either correct or close enough to fix), and 1% of the time it seems to actually fill in a substantial portion in a roughly acceptable way.

It's exceedingly frustrating and annoying, but not sure I can call it a net loss in time.

So reviewing the proposal for relevance and cut off and edits adds time to my workflow. Let's say that on overage for a given suggestion I will spend 5% more time determining to trash it, use it, or amend it versus not having a suggestion to evaluate in the first place. If the 20% useful time is 500% faster for those scenarios, then I come out ahead overall, though I'm annoyed 80% of the time. My guess as to whether the suggestion is even worth looking at improves, if I'm filling in a pretty boilerplate thing (e.g. taking some variables and starting to write out argument parsing), then it has a high chance of a substantial match. If I'm doing something even vaguely esoteric, I just ignore the suggestions popping up.

However, the 20% is a problem still since I'm maybe too lazy and complacent and spending the 100 milliseconds glancing at one word that looks right in review will sometimes fail me compared to spending 2-3 seconds having to type that same word out by hand.

That 20% success rate allowing for me to fix it up and dispose of most of it works for code completion, but prompt driven tasks seem to be so much worse for me that it is hard to imagine it to be better than the trouble it brings.

[–] jj4211@lemmy.world 23 points 2 days ago (2 children)

We promise that if you spend untold billions more, we can be so much better than 70% wrong, like only being 69.9% wrong.

[–] jj4211@lemmy.world 1 points 2 days ago

As someone who tries to keep the vague number in mind, it would be strange to me as well, but I suspect a large number of people don't really try to keep even the vague numbers in mind about how many people are about or how many people realistically could reside in a place like NYC.

They track the rough oversimplifications. Like "barely anyone lives in the middle of the country", and every TV show they see in the US either has a bunch of background people in NYC or LA, or is in the middle of nowhere with a town seemingly made up of mere dozens of people. They might know that "millions" live in the US and also, "millions" live in NYC, so same "ballpark" if they aren't keeping track of the specifics. They'd probably believe 10 million in NYC and 50 million nationwide.

This is presuming they bother to follow through on the specific math rather than merely roughly throwing out a percentage.

[–] jj4211@lemmy.world 3 points 2 days ago (2 children)

One thing I wonder is how seriously people take the flood warnings.

Most of the time if it is raining at all, I get the various flood warnings. I could imagine people underestimating those.

I recall quite a bit being made of how overtly grim, specific, and certain the Katrina warning was and how that may have helped set it apart from the usual "warning"

[–] jj4211@lemmy.world 1 points 2 days ago

My experience is that a limit of 256 means they probably are willing to allocate up to 24 bits to send the value over the network:

0x323536

People seem to love to pass around their numbers as JSON or similar.

[–] jj4211@lemmy.world 2 points 2 days ago

Even if true, 256 would be a waste of the range. 255 would make sense if trying to stay in one byte, using a whole different data type to get one extra bit just to hold 256 instead of saying "screw it, let's go to 511" even while using other bits.

It's just a very weird thing to do to pick 256 as a value limit back in those days (also oddly specific now, but for different reasons)

view more: next ›