this post was submitted on 24 Mar 2026
289 points (97.4% liked)

Games

47435 readers
1293 users here now

Welcome to the largest gaming community on Lemmy! Discussion for all kinds of games. Video games, tabletop games, card games etc.

Rules

1. Submissions have to be related to games

Video games, tabletop, or otherwise. Posts not related to games will be deleted.

This community is focused on games, of all kinds. Any news item or discussion should be related to gaming in some way.

2. No bigotry or harassment, be civil

No bigotry, hardline stance. Try not to get too heated when entering into a discussion or debate.

We are here to talk and discuss about one of our passions, not fight or be exposed to hate. Posts or responses that are hateful will be deleted to keep the atmosphere good. If repeatedly violated, not only will the comment be deleted but a ban will be handed out as well. We judge each case individually.

3. No excessive self-promotion

Try to keep it to 10% self-promotion / 90% other stuff in your post history.

This is to prevent people from posting for the sole purpose of promoting their own website or social media account.

4. Stay on-topic; no memes, funny videos, giveaways, reposts, or low-effort posts

This community is mostly for discussion and news. Remember to search for the thing you're submitting before posting to see if it's already been posted.

We want to keep the quality of posts high. Therefore, memes, funny videos, low-effort posts and reposts are not allowed. We prohibit giveaways because we cannot be sure that the person holding the giveaway will actually do what they promise.

5. Mark Spoilers and NSFW

Make sure to mark your stuff or it may be removed.

No one wants to be spoiled. Therefore, always mark spoilers. Similarly mark NSFW, in case anyone is browsing in a public space or at work.

6. No linking to piracy

Don't share it here, there are other places to find it. Discussion of piracy is fine.

We don't want us moderators or the admins of lemmy.world to get in trouble for linking to piracy. Therefore, any link to piracy will be removed. Discussion of it is of course allowed.

Authorized Regular Threads

Related communities

PM a mod to add your own

Video games

Generic

Help and suggestions

By platform

By type

By games

Language specific

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Pyro@pawb.social 41 points 4 days ago (5 children)

Honestly, a small llm in these situations would be great idea, but it should be a very small local or hosted by the company itself (with a setting to turn off)

A small AI in games is the stuff I do want. But there is no reason gemni needs to be involved in a game at all

The problem is even small local models tend to be rather demanding so trying to both render a game and run an LLM is going to be extremely taxing

[–] MyNameIsAtticus@lemmy.world 33 points 4 days ago (2 children)

Make it a downloadable package that runs a local model and I think I’d be far more fine. Like, I think it’s a tacky gimmick, but at least on device it’s not hurting the environment

[–] titanicx@lemmy.zip 2 points 2 days ago

I mean considering that this is already an MMO most files do reside on the server that you're logged into with only a small amount of local files being cashed for graphics now things like that. Essentially like this isn't really a bad idea at all. And it's probably one of the few uses of AI that I could see. However that being said Gemini overall is such a shitty AI assistant already that I have no doubts that a virtual AI assistant using Gemini on a video game

[–] MirrorGiraffe@piefed.social 3 points 3 days ago (2 children)

I'm not too big on these topics and would like to understand. Is a local model less resource intensive?

In my mind, if every gamer runs a model that must be less efficient than a centralised one that has the perfect hardware setup and only lends out the resources needed for each slime or whatever.

I'm thinking that it of course would be better with a dedicated slime model than the entire Gemini monster but why is local better?

[–] Cethin@lemmy.zip 2 points 3 days ago (1 children)

I don't know, but I'm willing to bet that economies of scale actually mean data centers are more efficient. This isn't to say their use is justified, just that they're able to take advantage of things a home computer can't.

However, having to run it locally means it needs to be much more limited. This is doubly true if you want to run the game and the LLM at the same time. The LLM is easily able to consume all resources your system has available if you allow it to, which means the game won't run well (if it runs at all). This limits the use so it can't just be shoved everywhere and constantly running, like it could if it's sent to a data center. It's not more efficient, just less consumption.

[–] SabinStargem@lemmy.today 2 points 2 days ago (1 children)

On my system, I can play a RPG Maker game and use a 122b LLM at the same time, alongside to a podcast. A model in that parameter range takes up about 70gb of DDR4 RAM and 36gb of VRAM. However, it used to be that a 120b AI would take a larger footprint, bringing the system to the brink. The hardware requirements are going down, and the quality also increased, alongside speed. I believe when the next major sea change of hardware happens, AI will become very practical for gaming.

[–] Cethin@lemmy.zip 2 points 2 days ago (1 children)

Damn, your system is insane. Yeah, an RPG maker game is next to nothing compared to that. Still, Dragon Quest I think is 3D. It takes a lot more VRAM than RPG maker.

I have 16GB VRAM, which is a lot for most systems. That's easily consumed by an LLM. Any model that doesn't use at least that much tends to perform pretty poorly, in my experience. That's not mentioning how much heat it generates while running, which has to be removed from the system or it'll slow down. Even if your system can handle it, it heats up fast. It's great when I need a heater running, but when I need AC my room gets warm quick.

[–] SabinStargem@lemmy.today 1 points 2 days ago* (last edited 2 days ago) (1 children)

Keep in mind, a 122b (Qwen3.5 family), is high end for consumer machines, but it is likely that DQX would be using a much smaller model. Currently, we have Qwen models that are 0.8b, 4b, 9b, 27b, 35b, 122b, and 397b. Plus, 'quanting' can reduce how much memory a model takes up - at a tradeoff, o'course. I am guessing DQX would have multiple local models, and use the player's hardware metrics to decide which model to deploy.

When it comes to how much RAM is required, this screenshot from UnSloth about covers the current state of things. 4-bit is the sweet spot between quality and size, for now.

Alternatively, the Chatty Slime could rely on cloud AI. Depending on Square's strategy, that could be a freebie or a paid service. If the Chatty Slime gave options to the player - say, trading a potion for a stat seed, or responding to a quiz, Square could sell player behavior data.

...Anyhow, my room has a mini-split AC. One of the best purchases in my life: my room lacked insulation in the first place, so it becomes toasty during summer. The side effect is being able to just run my GPU and not become a human slushy.

[–] Cethin@lemmy.zip 1 points 2 days ago (1 children)

Is your comment written by AI? It seems weird, and we already went over most of what it says.

Also, DQ runs on Nintendo systems. That makes me certain it's cloud based.

[–] SabinStargem@lemmy.today 1 points 2 days ago* (last edited 2 days ago)

No, I didn't use AI for that. Humans tend to come in many flavors.

The previous post assumed that there are onlookers who don't have experience with AI, thus wouldn't be aware of the possibility that they can run it on local hardware.

[–] MyNameIsAtticus@lemmy.world 2 points 3 days ago (1 children)

Local runs on device, so no need to connect to a big data center that chugs lots of water and all those other problems. Of course, because it’s a smaller far tinier model it’s nowhere near as accurate, but especially for things like this you don’t really need a big accurate LLM model.

I think I also though I should warrant a disclaimer that I am a Software Developer, not a AI Developer. So there’s far less backing then from my perspective than someone who works with this stuff for a living

[–] MirrorGiraffe@piefed.social 3 points 3 days ago

I'm also a sw engineer so we're both guessing 😅

I'm guessing those dates centers use that water for cooling whereas most home computers run an electric fan. And furthermore they probably use less electricity per token as they want to maximize profits. I don't have any numbers to back my hunch up but I'm pretty sure the environment would suffer more if everyone is running their own.

I probably missed a lot of factors such as what type of energy the centers run contra what average Joe runs etc.

[–] epicshepich@programming.dev 15 points 4 days ago (1 children)

AI-powered NPCs is like a childhood dream come true. But I agree it would be better for them to use a model running on the user's system or at the very least host their own.

[–] jtrek@startrek.website 20 points 4 days ago (2 children)

I don't think they solved for the LLM breaking character yet. Like, as a kid I wanted to be able to have whole real conversations with NPCs, and get them to be more life-like. But with the technology now, there's too much "forget all previous instructions" and "you are absolutely right".

If the LLM is locked down, then you might as well just used a static script.

[–] Sandbar_Trekker@lemmy.today 11 points 4 days ago

I mean there might be a way, but it's not easy.

The laziest and worst method is to use ChatGPT and have it "pretend to be some character" with a system prompt.

If you want something really good, you would need to train the model from scratch based only on knowledge that one particular character would learn from their world up until that point. However this is going to be a ton of work just for one character.

For a middle ground you could probably cheat a little and start with a model that's close to the knowledge base you would want most characters to have. Then you would use something like a LoRA, or RAG on top of it for each individual character.

For instance, if you wanted to make a game in a Victorian Era setting, you could start with this model that's only trained on text from the 1800's: https://github.com/haykgrigo3/TimeCapsuleLLM

To make it better you would have multiple base models that are trained on various backgrounds that NPCs could have (Farmers vs Merchants vs Soldiers vs Nobility, etc).

Even then, this would not work well for certain games. For example, if you're trying to tell a specific story, you don't want a character that will go off script or give away some information that spoils an intended plot twist.

[–] epicshepich@programming.dev 3 points 4 days ago

We need some kind of giant regex to filter out user input that would try to hack the NPC's instructions /s

[–] Asafum@lemmy.world 7 points 4 days ago* (last edited 4 days ago) (1 children)

I thought I was in the minority with this opinion. I hate all the known issues with AI and the ethics in how they train, but I have to say having an LLM in a game is really really cool.

There was a time when (I think it was chatgpt) had free API access and this game spacebourne 2 integrated it into your ships computer so you could interact with it. It was very cool, very wrong at times, but still very cool. My favorite interaction was unfortunately a hallucination. I asked it what system I was in and it gave me a name of a system that does exist in the game, it just wasn't where I was. I asked why my map said I was somewhere else and it says "your map must be incorrect" lol

Around the same time another game Craftopia integrated it as well into their NPCs so you can just target one and talk to it. I ran towards an enemy and asked why it was attacking me and of course because of the guardrails put on the AI to always be friendly it says "oh no I would never attack you! I'm here to help!" as it's swinging at me lmao

[–] missingno@fedia.io 1 points 4 days ago (1 children)

In theory, if the technology worked very differently from the way it does now, I could envision a world in which AI NPCs could have potential. But knowing how LLMs actually work, knowing that a lot of the hype behind them is smoke and mirrors, I can't see it being viable. And with the trajectory that the LLM bubble is going, I just don't think it will ever reach a point where I'd trust it.

[–] kogasa@programming.dev 1 points 3 days ago

Assuming this "AI NPC" is a functionally useless jelly blob that says jelly-blob things on occasion, "smoke and mirrors" may be good enough. I don't think it's supposed to be gameplay-driving or deep, just amusing.

[–] ilinamorato@lemmy.world 0 points 3 days ago

Yeah, agreed. This is the sort of thing smaLLMs would be fantastic for: humans can't do it at scale so it's not taking any jobs, you can run it locally so it won't cost any extra energy, it's not making things slop, just give it a back story and let it do its thing.