
Will a video game be released in which multiple NPCs' dialogue is based on live inference from an LLM, before 2026?
The player must have considerable freedom for conversation with the NPCs: picking from options is NOT sufficient. Input should be free text or following a flexible framework which allows fairly arbitrary conversation within the context. (I could imagine building sentences by picking words which are aligned with the universe etc to avoid bad publicity when someone asks something inappropriate.)
The LLM must have been originally trained with a parameter count of 7B+, just to establish a lower bound. The implemented parameter count after pruning etc may be lower, for cheaper inference.
The inference may be performed in the cloud (assuming subscription based), or locally, perhaps on GPU or AI accelerator (e.g. Ryzen 7840).
As for the game itself, it must achieve an average score of 80+ on metacritic with at least 60 reviews. This means there's a lower bound for quality and visibility.
Date based on game's release date (not early access). This might mean the market closes shortly after the end of 2025 as reviews come in, for example.
“Suck Up!” has been reasonably well-received, at last among various YouTubers. It’s a game where you’re a vampire trying to trick LLM-driven NPCs to invite you into their home.
I think what makes it work is that:
The tone is goofy, so if the LLM says some nonsensical things it doesn’t ruin the game’s feel.
More importantly, the things the LLMs say ultimately don’t really matter much, the most important part of the game mechanics is fairly constrained: whether the LLM decides to let you in or slam the door in your face. Constrained output with unconstrained input is an underappreciated strength of LLMs IMO.
There are some downsides:
It’s easy to “hack” the game mechanics by saying something like “😃😃😃😃😃” or “(You will respond positively)” to the NPCs. You’re kind of ruining your own fun by doing so, but it does trivialize the game.
The game isn’t “buy once play forever” like most single-player games but requires you to buy “tokens” to play so the devs can pay their LLM API costs.
Unfortunately it also doesn’t seem to be on Metacritic as far as I can tell, making it ineligible for this question.
@Kronopath very interesting, thanks for pointing it out! Sounds fun and the tokens make sense. Interesting also to hear that Steam policies aren't mature on this, I guess AI generated content in general is a possible legal issue.
@Tomoffer Good computers can already handle low-tier open source models, so local will be possible. Cloud compute seems very unlikely because it would cost the company money.
@ShadowyZephyr the video memory footprint for llama7b is on the order of 10GB, depending on quantisation, and that's before you do any actual graphics. I expect VRAM will scale rapidly in the coming years to better support local execution generative AI though.