brucethemoose

joined 1 year ago
[–] brucethemoose@lemmy.world 1 points 7 minutes ago* (last edited 3 minutes ago)

Musk has quite a “tech bro” following (which we don’t see because we don’t live and breathe on Twitter and such), and that group wields enormous psychological influence over the population.

Seems unlikely, but If Musk aligns himself with Peter Theil, Zuckerberg, maybe Google and such more closely, that’s an extremely dangerous platform for Trump. They can sap power from MAGA (and anyone else) with the flip of a switch.

There’s quite a fundamental incompatibility between tech oligarchs and the red meat MAGA base, too, as is already being exposed. It’s going to get much more stark over the next few years.

[–] brucethemoose@lemmy.world 62 points 22 hours ago* (last edited 22 hours ago) (6 children)

Narcissists hate being ignored or called unimportant. Trump flippantly dismissing him as “nuts” and moving on is the ultimate insult.

I’m sure Musk has an army reining him in, but that’s legitimately hard for Musk to ignore.

[–] brucethemoose@lemmy.world 2 points 1 day ago* (last edited 1 day ago) (1 children)

One one more thing, I saw you mention context management.

Mistral (24B) models are really bad at long context, but this is not always the case. I find that Qwen 32B and Gemma 27B are solid at 32K (which is a huge body of text) and (with the right backend settings) you can easily run either at 64K with very minimal vram overhead.

Specifically, run Gemma with the latest llama.cpp server and comment (where it will automatically use sliding window attention as of like yesterday), or Qwen (and most other models) with exllamav2 or exllamav3, which quantizes the kv cache down to Q4 very efficiently.

This way you don’t need to manage context: you can feed the LLM the whole adventure so it doesn’t forget anything, and streaming responses will be instance since it’s always cached.

[–] brucethemoose@lemmy.world 1 points 1 day ago

Oh, one thing about ST specifically: its default sampling presets are catastrophic last I checked. Like, they’re designed for ancient models, and while I have nothing against the UI it is kinda from a different era.

For Gemma and Qwen, I’ve been using like 0.2-0.7 temp, at least 0.05 MinP, 1.01 rep penalty (not something insane like 1.1) and maybe 0.3-ish dry, though like you said dry/xtc can really mess up some tasks.

[–] brucethemoose@lemmy.world 2 points 1 day ago* (last edited 1 day ago) (4 children)

Also, another suggestion would be to be careful with your sampling. Use a low temperature and high MinP for queries involving rules, higher temperature (+ samplers like DRY) when you're trying to tease out interesting ideas.

I would even suggest an alt front end like mikupad that exposes token probabilities, so you can go to any point in the reply and look through every “idea” the LLM had internally (and regen from that point of you wish”). It’s also good for debugging sampling issues when you have an incorrect answer (as sometimes the LLM gets it right, but bad sampling parameters choose a bad answer).

[–] brucethemoose@lemmy.world 1 points 1 day ago

As long as it supports network inference between machines with heterogeneous cards, it would work for what I have in mind.

It probably doesn’t, heh, especially non Nvidia cards. But the middle layer may work with some generic OpenAI backend like the llama.cpp server.

[–] brucethemoose@lemmy.world 16 points 1 day ago

Yeah, most predatory apps are almost like cheap ripoffs of refined system casinos got down to a science.

[–] brucethemoose@lemmy.world 35 points 1 day ago* (last edited 1 day ago) (15 children)

Both can be true.

It can be true that the FDA was corrupted/captured to some extent and needs more 'skeptial' and less-industry-friendly leadership. At the same time, skepticism in science is not the answer.

This is my dillema with MAGA. Many of the issues they tackle are spot on, even if people don't like to hear that. They're often right, even when the proposed solutions are wrong and damaging. I think this a lot when I hear RFK speak, nodding my head at the first assertion then grinding my teeth as he goes on.

[–] brucethemoose@lemmy.world 1 points 1 day ago

The budget for marketing has doubled the cost of the entire previous game. Does anyone need ads for GTA6? Wouldn’t just having the devs do livestreams of them playing the game and discussing the tech involved with making GTA6 not create enough hype? Does there even need to be additional hype created?

There is a bit of an "arms race," where other games/entertainment could steal GTA's engagement. Eyeball time is finite, and to quote a paper, "attention is all you need."

You aren't wrong though. Spending so much seems insane when "guerrilla marketing" for such a famous IP would go a long way. I guess part of it is "the line must go up" mentality, where sales must increase dramatically next quarter even if that costs a boatload of marketing to achieve.

[–] brucethemoose@lemmy.world 2 points 2 days ago (7 children)

Late to the post, but look into SGLang, OP!

In a nutshell, it’s a framework for letting LLMs “fill in blanks” instead of generating entire replies, so you could script in rules as part of the responses as structure for it to grab onto. It’s all locally runnable (with the right hardware, unfortunately).

Also, there are some newer, less sycophantic DM specific models. I can look around if you want.

[–] brucethemoose@lemmy.world 6 points 2 days ago* (last edited 2 days ago)

This hurts as a Texan, but also rings true. I used to think we’re more “independent” minded than the South (as I have some scary Southern family), but every day since 2016 has opened my eyes more.

The murder made me think about how much I heard “fag,” racial slurs and worse as a kid, and how many stayed in the closet out of fear.

If you come, come to Austin! It’s better. Avoid Dallas, it sucks.

[–] brucethemoose@lemmy.world 4 points 2 days ago

Is it a virus that affects the brain?

Yes! It’s called engagement optimization. And the worlds collective ignorance “don’t feed the trolls.”

 

In a nutshell, he’s allegedly frustrated by too few policies favorable to him.

 
  • The IDF is planning to displace close to 2 million Palestinians to the Rafah area, where compounds for the delivery of humanitarian aid are being built.
  • The compounds are to be managed by a new international foundation and private U.S. companies, though it's unclear how the plan will function after the UN and all aid organizations announced they won't take part
 

Qwen3 was apparently posted early, then quickly pulled from HuggingFace and Modelscope. The large ones are MoEs, per screenshots from Reddit:

screenshots

Including a 235B/22B active and a 30B/3B active.

Context appears to 'only' be 32K unfortunately: https://huggingface.co/qingy2024/Qwen3-0.6B/blob/main/config_4b.json

But its possible they're still training them to 256K:

from reddit

Take it all with a grain of salt, configs could change with the official release, but it appears it is happening today.

 

This is one of the "smartest" models you can fit on a 24GB GPU now, with no offloading and very little quantization loss. It feels big and insightful, like a better (albeit dry) Llama 3.3 70B with thinking, and with more STEM world knowledge than QwQ 32B, but comfortably fits thanks the new exl3 quantization!

Quantization Loss

You need to use a backend that support exl3, like (at the moment) text-gen-web-ui or (soon) TabbyAPI.

 

"It makes me think that maybe he [Putin] doesn't want to stop the war, he's just tapping me along, and has to be dealt with differently, through 'Banking' or 'Secondary Sanctions?' Too many people are dying!!!", Trump wrote.

 

The U.S. expects Ukraine's response Wednesday to a peace framework that includes U.S. recognition of Crimea as part of Russia and unofficial recognition of Russian control of nearly all areas occupied since the 2022 invasion, sources with direct knowledge of the proposal tell Axios.

What Russia gets under Trump's proposal:

  • "De jure" U.S. recognition of Russian control in Crimea.
  • "De-facto recognition" of the Russia's occupation of nearly all of Luhansk oblast and the occupied portions of Donetsk, Kherson and Zaporizhzhia.
  • A promise that Ukraine will not become a member of NATO. The text notes that Ukraine could become part of the European Union.
  • The lifting of sanctions imposed since 2014.
  • Enhanced economic cooperation with the U.S., particularly in the energy and industrial sectors.

What Ukraine gets under Trump's proposal:

  • "A robust security guarantee" involving an ad hoc group of European countries and potentially also like-minded non-European countries. The document is vague in terms of how this peacekeeping operation would function and does not mention any U.S. participation.
  • The return of the small part of Kharkiv oblast Russia has occupied.
  • Unimpeded passage of the Dnieper River, which runs along the front line in parts of southern Ukraine.
  • Compensation and assistance for rebuilding, though the document does not say where the funding will come from.

Whole article is worth a read, as it’s quite short/dense as Axios usually is. For those outside the US, this is an outlet that’s been well sourced in Washington for years.

 

Seems there's not a lot of talk about relatively unknown finetunes these days, so I'll start posting more!

Openbuddy's been on my radar, but this one is very interesting: QwQ 32B, post-trained on openbuddy's dataset, apparently with QAT applied (though it's kinda unclear) and context-extended. Observations:

  • Quantized with exllamav2, it seems to show lower distortion levels than nomal QwQ. Its works conspicuously well at 4.0bpw and 3.5bpw.

  • Seems good at long context. Have not tested 200K, but it's quite excellent in the 64K range.

  • Works fine in English.

  • The chat template is funky. It seems to mix up the and <|think|> tags in particular (why don't they just use ChatML?), and needs some wrangling with your own template.

  • Seems smart, can't say if it's better or worse than QwQ yet, other than it doesn't seem to "suffer" below 3.75bpw like QwQ does.

Also, I reposted this from /r/locallama, as I feel the community generally should going forward. With its spirit, it seems like we should be on Lemmy instead?

 

So I had a clip I wanted to upload to a lemmy comment:

  • Tried it as an (avc) mp4... Failed.
  • OK, too big? I shrink it to 2MB, then 1MB. Failed.
  • VP9 Webm maybe? 2MB, 1MB, failed. AV1? Failed.
  • OK, fine, no video. Lets try an animated AVIF. Failed. It seems lemmy doesn't even take static AVIF images
  • WebP animation then... Failed. Animated PNG, failed.

End result, I have to burden the server with a massive, crappy looking GIF after trying a dozen formats. With all due respect, this is worse than some aging service like Reddit that doesn't support new media formats.

For reference, I'm using the web interface. Is this just a format restriction of lemmy.world, or an underlying software support issue?

 

53% of Americans approve of Trump so far, according to a newly released CBS News/YouGov poll conducted Feb. 5 to 7, while 47% disapproved.

A large majority, 70%, said he was doing what he promised in the campaign, per the poll that was released on Sunday.

Yes, but: 66% said he was not focusing enough on lowering prices, a key campaign trail promise that propelled Trump to the White House.

44% of Republicans said Musk and DOGE should have "some" influence, while just 13% of Democrats agreed.

 

Here's the Meta formula:

  • Put a Trump friend on your board (Ultimate Fighting Championship CEO Dana White).
  • Promote a prominent Republican as your chief global affairs officer (Joel Kaplan, succeeding liberal-friendly Nick Clegg, president of global affairs).
  • Align your philosophy with Trump's on a big-ticket public issue (free speech over fact-checking).
  • Announce your philosophical change on Fox News, hoping Trump is watching. In this case, he was. "Meta, Facebook, I think they've come a long way," Trump said at a Mar-a-Lago news conference, adding of Kaplan's appearance on the "Fox and Friends" curvy couch: "The man was very impressive."
  • Take a big public stand on a favorite issue for Trump and MAGA (rolling back DEI programs).
  • Amplify that stand in an interview with Fox News Digital. (Kaplan again!)
  • Go on Joe Rogan's podcast and blast President Biden for censorship.
 

Taboola's data, shared exclusively with Axios, shows Musk has outpaced his closest peers — Jeff Bezos and Mark Zuckerberg — for years, but the gap widened dramatically in 2024.

The spam is already exponential. :(

372
submitted 5 months ago* (last edited 5 months ago) by brucethemoose@lemmy.world to c/politics@lemmy.world
 

Reality check: Trump pledged to end the program in 2016.

Called it. When push comes to shove, Trump is always going to side with the ultra-rich.

view more: next ›