As if EA didn’t already make bland, derivative games…

  • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 ℹ️
    link
    fedilink
    English
    19
    edit-2
    6 months ago

    I unironically am; but I don’t think EA is gonna use it the way I would actually like to see it used in a video game. They hear “I want AI dialogue” and will use AI to write (and probably code) the game, instead of having it generated real time so you can communicate with the NPCs as if they were actually there.

    • @[email protected]
      link
      fedilink
      246 months ago

      The problem is that you can’t really control what AI spits out. May fit perfectly into the story, or it may be immersion breaking nonsense that doesn’t even fit into the narrative. What if a character suddenly makes a promise or tells you a key plotline point that it has just made up? I, for one, prefer games to be handcrafted to deliver a quality reliable experience instead of being a coinflip.

      • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 ℹ️
        link
        fedilink
        English
        -8
        edit-2
        6 months ago

        You can control what it spits out, though. They already do somewhat.

        Edit: Gonna go out on a limb and assume most of you haven’t actually played any of the projects currently doing this. Or mess with chatbots at all.

        • @[email protected]
          link
          fedilink
          196 months ago

          Somewhat is key. You can try to guide it in a direction, but that’s it. Also, as a player, you can never be sure if the dialogue is meaningful or not. Does it reveal something about the plot? Is it a key information about the character? Is it just hallucinated gibberish to fill the space?

          • @[email protected]
            link
            fedilink
            56 months ago

            Besides, LLMs struggle with retaining contextual information for long and they’re pretty dang resource hungry. Expect a game with LLM-driven dialogue to reserve several gigs of VRAM and a fair chunk of GPU processing power solely for that.

            And then you still get characters who hallucinate plot points or suddenly speak gibberish.

        • MentalEdge
          link
          fedilink
          3
          edit-2
          6 months ago

          You really can’t.

          You can run checks and fence it in with traditional software, you can train it more narrowly…

          I haven’t seen anything that suggests AI hallucinations are actually a solvable problem, because they stem from the fact that these models don’t actually think, or know anything.

          They’re only useful when their output is vetted before use, because training a model that gets things 100% right 100% of the time, is like capturing lightning in a bottle.

          It’s the 90/90 problem. Except with AI it’s looking more and more like a 90/99.99999999 problem.