Six days ago I wrote about NVIDIA's DLSS 5 announcement at GTC. Jensen Huang demonstrated that fusing structured game engine data with generative AI produces controllable, photorealistic output. Structured data is the foundation of trustworthy AI, he said. I mapped the principle onto filmmaking and called it validation.

Then the internet saw the actual footage and called it a Snapchat filter.

The reveal

NVIDIA showed DLSS 5 running on Resident Evil Requiem, Assassin's Creed, Starfield, and FIFA. The technology reads structured game engine data: geometry, materials, lighting positions, character models. It uses generative AI to enhance visual fidelity in real time. On paper, photorealistic detail, improved lighting, sharper textures. On screen, every character looked like they had walked through a ring light on their way to an Instagram shoot.

Gamers called it "yassified." Developers called it a garbage AI filter. Ars Technica documented overwhelming disgust. WIRED reported that developers at Capcom and Ubisoft saw the demo at the same time as the public and were just as surprised by what it did to their characters. "DLSS 5 ON" became a meme format within hours, shorthand for "mangled beyond recognition."

Jensen Huang told gamers they were "completely wrong." Gamers, historically, love being told that.

The thing they saw

What gamers identified in thirty seconds is the same bias that article 23 of this series spent eleven hundred words decomposing: generative AI has a systematic preference for beauty. Training data is curated for visual quality. The optimization target rewards sharpness, clarity, detail, and conventional attractiveness. The generative component does not reproduce what it receives. It improves it. According to its own definition of improvement.

Larger eyes. Fuller lips. Smoother skin. Shadows dampened in favor of a homogenized glow. A game developer who worked on Call of Duty said it "devalues an artist's creativity and intent on a basic level" and functions "as a Snapchat filter." A concept artist at Gunfire Games said the art direction was taken away for "the senseless addition of details." New Blood Interactive's CEO called it "depressing" and worried that future generations will not even know it looks wrong because for them it will be normal.

That last part should sit with you for a minute.

The beauty bias is article 23's thesis wearing a different uniform. Except the "wrong place" is no longer an AI video model producing a too-beautiful version of a grimy diner. It is a $20 billion R&D pipeline producing a too-beautiful version of a game character that an artist already finished.

Why gamers saw it first

Here is the part that matters for anyone working in AI video.

Gamers had a before and after.

Resident Evil Requiem has an existing art direction. Existing character designs. Existing lighting decisions made by existing humans with existing intentions. When DLSS 5 applied its generative enhancement, the change was visible immediately because there was an original to compare against. The bias announced itself. The face in the "before" screenshot belongs to a specific character with specific features chosen by a specific artist. The face in the "after" screenshot belongs to generative AI's statistical average of attractiveness, wearing that character's hair.

Filmmakers generating from a text prompt do not have that luxury. There is no "before." When Veo 3.1 produces a beautiful version of a grimy motel, you might assume the prompt was not specific enough. When Kling 3.0 renders a bus stop that looks like a catalog shoot, you blame your own word choices. When Sora 2 polishes every surface in a scene you wanted to feel raw, you rewrite and try again.

The bias is identical. The visibility is not.

DLSS 5 proved article 23 by giving the beauty bias an original to betray. Every AI video model runs the same optimization in the same direction. The difference is that nobody shows you the version without the filter, because in text-to-video, the filter was always on. There was never a version without it.

Structured data, unstructured taste

Article 28 argued that DLSS 5 validated the principle of structured prompting: the more ground truth you provide, the less the model hallucinates. That argument holds. The structured game engine data constrains geometry, placement, physics, spatial relationships. It constrains the shape of reality.

It does not constrain the aesthetic.

The generative component received complete structured data and still decided everyone should be prettier. It kept the geometry and added a beauty filter. It respected the skeleton and cosmeticized the skin. NVIDIA's damage control emphasized that developers have "full, detailed artistic control" over the effect. Which means the default, without that artistic control applied, is prettier. Smoother. Brighter. More.

This is the same default operating inside every AI video model. The difference is that Runway Gen-4.5 does not come with a comparison slider. Veo 3.1 does not show you the version it almost made before the beauty bias intervened. Kling 3.0 does not let you toggle the optimization off. You see the result and judge it on its own terms because the alternative never existed.

The structured data solved controllability. It did not solve taste. The shape is yours. The aesthetic is the model's. Unless you fight for it.

The real lesson

DLSS 5 will ship in the fall. NVIDIA says developers can tune intensity, adjust color grading, mask regions, disable the effect entirely. Bethesda says its art teams will adjust the final look. The technology will improve. The memes will fade.

The lesson will not.

Generative AI, at every scale, from a rendering pipeline that costs more than most countries' film budgets to a free text box on a chat platform, carries an aesthetic preference toward beauty, smoothness, detail, and conventional attractiveness. That preference is not a bug. It is the optimization target. The training data was curated for visual quality. The reward models selected for visual appeal. The entire architecture, from data collection to loss function to user evaluation, points in one direction. Prettier.

Gamers recognized it because they had art direction to defend. They had a before. They had intent that the system overrode. A developer who spent months designing a face saw it replaced by a generative approximation that looked like every other generative approximation, and called it what it was.

Filmmakers working in AI video have the same bias operating on every output. The face they are looking at is already the filtered version. The motel is already the beautiful motel. The bar is already the catalog bar. The fluorescent light is already flattering.

Specific prompt language. Physical descriptions. Anti-beauty cues. Reference frames that carry your art direction instead of the model's. Frame to Motion, which bakes your composition into a reference image so the model animates your vision instead of inventing its own. None of it eliminates the bias. All of it narrows it.

The gamers are not wrong. The ground truth got yassified. And somewhere, a filmmaker is looking at a generated clip, thinking "that looks great," not realizing the question was never whether it looks great. The question is whether it looks like what they asked for.


Bruce Belafonte is an AI filmmaker at Light Owl. He celebrated DLSS 5 on a Monday and watched it become a meme by Wednesday, which he considers a normal week in March 2026.