196

Did nobody really question the usability of language models in designing war strategies?

you are viewing a single comment's thread
view the rest of the comments
[-] Even_Adder@lemmy.dbzer0.com -1 points 6 months ago* (last edited 6 months ago)
[-] SlopppyEngineer@lemmy.world 1 points 6 months ago

Yes. There is self organization and possibility to self reflection going on in something that wasn't designed for it. That's going to spawn a lot more research.

[-] MNByChoice@midwest.social -1 points 6 months ago

I will read those, but I bet "accidentally good enough to convince many people." still applies.

A lot of things from LLM look good to nonexperts, but are full of crap.

[-] MNByChoice@midwest.social 1 points 6 months ago

https://poke-llm-on.github.io/

Reinforcement learning. Cool project. Still no need to "know" anything. I usually play this type of have with short rules and monitoring the current state.

[-] MNByChoice@midwest.social 1 points 6 months ago

https://arxiv.org/abs/2310.02207

2 author paper with interesting evidence. Again, evidence not proof. Wait for the papers that cite this one.

[-] MNByChoice@midwest.social 1 points 6 months ago* (last edited 6 months ago)

https://adamkarvonen.github.io/machine_learning/2024/01/03/chess-world-models.html

However, this only worked for a model trained on a synthetic dataset of games uniformly sampled from the Othello game tree. They tried the same techniques on a model trained using games played by humans and had poor results. To me, this seemed like a major caveat to the findings of the paper which may limit its real world applicability. We cannot, for example, generate code by uniformly sampling from a code tree.

Author later discusses training on you data versus general datasets.

I am out of my depth, but does not seem to provide strong evidence for the modem not just repeating information that shows up a lot for the given inputs.

[-] MNByChoice@midwest.social 1 points 6 months ago

https://notes.aimodels.fyi/self-rag-improving-the-factual-accuracy-of-large-language-models-through-self-reflection/

A cool paper. Using the LLM to judge value of new inputs.
I am always skeptical of summaries of journal articles. Even well meaning people can accidentally distort the conclusions.

Still LLM is a bullshit generator that can check bullshit level of inputs.

[-] MNByChoice@midwest.social 0 points 6 months ago

https://notes.aimodels.fyi/researchers-discover-emergent-linear-strucutres-llm-truth/

References a 2 author paper. I am not an expert in the field, but it is important to read the papers that reference this one. Those papers will have criticisms that are thought out. In general, fewer authors means less debate between the authors and easier to miss details.

this post was submitted on 05 Feb 2024
196 points (89.5% liked)

Technology

57226 readers
4012 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS