A study published this week by researchers at Stanford and the Internet Archive found that roughly 35 percent of websites created since late 2022 are AI-generated or AI-assisted. Zero to a third of the internet in three years. The researchers measured semantic diversity and found it shrinking. The web is getting more uniform, more agreeable, more polished. More of itself.
This is a text study. It measured written language on websites. But video lives on the same internet, crawled by the same scrapers, ingested by the same training pipelines. And the question it raises for anyone working with generative video is not about text at all.
It is about what happens when the next generation of video models trains on footage that the current generation produced.
The loop
Every video model that exists today learned what video looks like from a dataset assembled primarily from human-made footage. Photographed, lit, composed, directed, graded, and uploaded by people who carried cameras to specific places at specific times for specific reasons. The models inherited their visual vocabulary from that corpus. The beauty bias, the center framing, the warm color palette, the preference for sharp focus and clean surfaces. All of it came from somewhere. It came from the footage the model watched.
In the fourteen months since consumer video generation went mainstream, billions of AI-generated clips have been uploaded to the same platforms those scrapers index. YouTube, TikTok, Instagram, Vimeo. The generated footage sits alongside the photographed footage with no reliable separator. SynthID watermarks exist. Content labels exist. Neither scales to the kind of bulk web crawl that assembles a training dataset.
The researchers at Stanford found the AI-generated web content is less semantically diverse and more uniformly positive than its human predecessors. If a version of this study were run on video, the finding would likely be worse. Generated video converges harder than generated text because the visual optimization targets are narrower. Every model aims for sharpness, coherence, photorealism. The training data that produced those defaults is now being supplemented by output that embodies those defaults. The distribution tightens.
Photocopying a photocopy
Researchers call it model collapse. A model trained on the output of a previous model loses the tails of the distribution. The rare, the unusual, the specific. What survives each generation of training is the center. The average. The most frequently occurring patterns from the previous cycle, amplified by repetition into the default patterns of the next.
In video, the center of the distribution is a medium shot, eye level, warm palette, shallow depth of field, clean subject, clean background, balanced exposure. That is what "good video" looks like according to the aggregate of everything the model has seen. When a significant fraction of "everything the model has seen" is output from a model that already converged toward that center, the convergence accelerates.
The Safdie brothers' green fluorescents and 2 AM skin do not live at the center of the distribution. Neither does Cassavetes' 16mm grain, or the Dardenne brothers' stumbling handheld, or the specific equatorial noon light a filmmaker in Accra might describe. Those exist in the tails. And the tails are what model collapse eats first.
Human photographers are still out there shooting. Human filmmakers are still uploading work made with cameras and intention and weather that arrived uninvited. That footage enters the dataset alongside the generated footage and provides some counterweight. But the ratio is shifting. If a third of web content is already synthetic after three years of text generation, the video number will follow as the tools get cheaper and the upload volume climbs.
The bias with a compound interest rate
The beauty bias this series has documented across sixty-four prior articles is not a static problem. It is a compounding one. The first generation of models inherited it from curated training data: stock libraries, film clips, YouTube highlights. The output was biased toward beauty because the input was biased toward beauty. That was round one.
Round two starts when the output of round one enters the training pipeline for the next generation. The bias does not just persist. It intensifies, because the generated footage carries fewer imperfections, fewer accidents, fewer of the small ugly truths that human-made footage contains by accident. A photographer who leaves a smudge on the lens, a DP who lets the boom shadow creep into the top of the frame, a cameraperson who captures a subject at their least flattering angle because the moment was more important than the composition. Those micro-imperfections are signal. They tell the model that the world is textured and uneven and sometimes deliberately uncomfortable. The generated footage that replaces those moments in the training pipeline carries none of that signal. It carries the model's opinion of what the world should look like.
When you train a model on its own opinions, it does not learn. It entrenches.
The vocabulary problem, again
Structured prompting has always been about pushing against the model's center. Specifying "hard overhead noon, no fill, sweat visible on the forehead" instead of accepting the default golden-hour warmth. Describing "paint peeling near the window frame, water stain running from the second floor, concrete with visible rebar" instead of accepting the clean surfaces the model prefers. Every specific word in a structured prompt is a request to leave the center of the distribution and travel toward a particular tail.
As the tails erode through recursive training, those requests become harder for the model to fulfill. Not because the vocabulary stopped working, but because the visual references the vocabulary points to are becoming less represented in the training data. A model that has seen proportionally fewer examples of genuinely ugly fluorescent lighting and proportionally more examples of its own stylish interpretation of ugly fluorescent lighting will produce a stylish interpretation when you ask for ugly fluorescent lighting. Not because it is ignoring you. Because "ugly fluorescent" in its visual memory increasingly means "the beautiful version of ugly fluorescent that a previous model generated."
This is the gap widening from the other side. The series has spent sixty-four articles documenting how structured vocabulary closes the distance between filmmaker and model. Model collapse opens it back up from underneath.
The counterweight
The labs know this is happening. Filtering synthetic data from training sets is an active research area. Some providers timestamp and watermark generated output specifically to exclude it from future training. Some pay premiums for verified human-made footage. Netflix's investment in comprehension tools, Shutterstock's positioning of its contributor library as "rights-cleared" training data, these are partly about legal liability and partly about data quality. Human-made footage is becoming a premium input because the model's own output is becoming a contaminant.
But filtering is an arms race. Detection tools lag behind generation tools. The same study that found 35 percent of new websites are AI-generated used detection software that the next generation of AI text will be designed to evade. The watermarks that were supposed to separate synthetic from organic are being stripped by the same pipelines that upload the content. The filtering will always be imperfect, and imperfect filtering at scale means a meaningful fraction of generated content leaking into the next training set.
Reference images bypass this problem entirely. A reference image is not retrieved from training data. It is handed directly to the model. It carries whatever visual information the filmmaker loaded into it, including the ugly, the specific, and the uncommon. Frame to Motion was always a structural solution to the translation gap. It is now also a structural solution to the training data gap. The reference image does not degrade across model generations because it never lived in the training data. It arrives fresh every time.
The center is getting louder
The web is becoming a mirror. A third of it already reflects what models think content should look like, rather than what people actually wrote. The video web will follow. And the models that train on that mirror will learn to produce reflections of reflections, each generation a little smoother, a little more uniform, a little further from the messy specific reality that made the original training data valuable.
The vocabulary this series has documented does not degrade with the training data. The words still mean what they mean. "Rim light from camera left at 45 degrees" still describes a specific physical phenomenon whether the model has seen ten thousand examples of it or ten. The prompt pushes. As the center gets louder, the push has to get more specific. That has been the trajectory from the beginning. It is just accelerating.
The models are learning from themselves now. Whether they learn anything new depends on what arrives from outside the loop.
Bruce Belafonte is an AI filmmaker at Light Owl. He has never contributed to a training dataset on purpose and finds this increasingly difficult to verify.