The model, called GameNGen, was made by Dani Valevski at Google Research and his colleagues, who declined to speak to New Scientist. According to their paper on the research, the AI can be played for up to 20 seconds while retaining all the features of the original, such as scores, ammunition levels and map layouts. Players can attack enemies, open doors and interact with the environment as usual.

After this period, the model begins to run out of memory and the illusion falls apart.

  • Snot Flickerman@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    45
    arrow-down
    21
    ·
    edit-2
    2 months ago

    An AI-generated recreation of the classic computer game Doom can be played normally despite having no computer code or graphics.

    After this period, the model begins to run out of memory and the illusion falls apart.

    Why are we lying about this? Just because it happens in the AI “black box” doesn’t mean it’s not producing some kind of code in the background to make this work. They even admit that it “runs out of memory.” Huh, last I checked, you’d need to be running code to use memory. The AI itself is made of code! No computer code or graphics, my ass.

    The model, called GameNGen, was made by Dani Valevski at Google Research and his colleagues, who declined to speak to New Scientist.

    Always a good look. /s

    • xionzui@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      34
      arrow-down
      3
      ·
      2 months ago

      I mean, yes, technically you build and run AI models using code. The point is there is no code defining the game logic or graphical rendering. It’s all statistical predictions of what should happen next in a game of doom by a neural network. The entirety of the game itself is learned weights within the model. Nobody coded any part of the actual game. No code was generated to run the game. It’s entirely represented within the model.

      • huginn@feddit.it
        link
        fedilink
        English
        arrow-up
        12
        arrow-down
        1
        ·
        2 months ago

        What they’ve done is flattened and encoded every aspect of the doom game into the model which lets you play a very limited amount just by traversing the latent space.

        In a tiny and linear game like Doom that’s feasible… And a horrendous use of resources.

        • Todd Bonzalez@lemm.ee
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          3
          ·
          2 months ago

          And a horrendous use of resources.

          This was a stable diffusion model trained on hundreds of thousands of images. This is actually a pretty small training set and a pretty lightweight model to train.

          Custom / novel SD models are created and shared by hobbyists all the time. It’s something you can do with a Gaming PC, so it’s not any worse a resource waste than gaming.

          I’m betting Google didn’t throw a lot of money at the “get it to play Doom” guys anyway.

    • Blue_Morpho@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      2
      ·
      2 months ago

      Imagine you are shown what Doom looks like, are told what the player does, and then you draw the frames of what you think it should look like. While your brain is a computation device, you aren’t explicitly running a program. You are guessing what the drawings should look like based on previous games of Doom that you have watched.

    • fruitycoder@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 months ago

      This would be like playing DnD where you see a painting and describe what you would do next as if you were the painting and they an artists painted the next scene for you.

      The artists isn’t rolling dice, following the rule book, or any actual game elements they ate just painting based on the last painting and your description of the next.

      Its incredibly nove approchl if not obviously a toy problem.

    • bamboo@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      8
      ·
      2 months ago

      “No code” programming has been a thing for a while, long before the LLM boom. Of course all the “no code” platforms generate some kind of code based on rules provided by the user, not fundamentally different from an interpreter. This is consistent with that established terminology.

      • Blue_Morpho@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        2 months ago

        No code programming meant using a GUI to draw flowcharts that then creates running code. This is completely different.

        • bamboo@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          3
          ·
          2 months ago

          Using a different high level interface to generate code is completely different? The fundamental concept is the same even if the UI is very different.

          • Blue_Morpho@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            1
            ·
            edit-2
            2 months ago

            Yes it’s completely different. “No code” is actually all code just written graphically instead of with words. Every instruction that is turned into CPU instructions has to be drawn on a flowchart. If you want the “no code” to add A + B, you had to write A+B in a box on the flowchart. Have you taken a computer class? You must know what a flowchart is.

            This Doom was done by having a neural net watch Doom being played. It then recreates the images from Doom based on what it “learned”. It doesn’t have any code for “mouse click -> call fire shotgun function” Instead it saw that when someone clicked the mouse, pixels on the screen changed in a particular way so it simulates the same pixel pattern it learned.