NVIDIA showed up at GDC this week and announced that ComfyUI, the node-based generation tool that looks like an electrician's nightmare, now has a simplified App View. Enter a prompt, adjust a few sliders, hit generate. On your own GPU. No API key. No queue. No per-clip invoice arriving at 3 AM.

Same week, they shipped RTX Video Super Resolution as a ComfyUI node. Generate at lower resolution, upscale to 4K in real time. On a consumer card. The RTX 5090 does this thirty times faster than popular open-source upscalers and uses a fraction of the memory.

ComfyUI also got NVFP4 support, which compresses model weights so they fit in less VRAM and run 2.5 times faster. The combination of all three announcements means a machine that costs four thousand dollars can now generate AI video, upscale it to 4K, and do both without an internet connection.

That is a genuine shift. Not a convenience. A relocation.

From rent to mortgage

For the past year, AI video generation lived in somebody else's building. Sora, Veo, Kling, Runway, Seedance, WAN, Grok Imagine: cloud services. You sent a prompt over the internet. A server somewhere rendered it. You paid per clip or per month or per credit. The generation happened elsewhere, controlled by someone else, on hardware you would never see.

An earlier piece in this series argued the generate button is a commodity. Different platforms wrapping the same APIs in different interfaces and charging different markups. CinePrompt's BYOK architecture stripped the middleman by letting users bring their own API keys and pay provider rates directly.

Local generation strips the provider too.

When the model runs on your hardware, the entire supply chain collapses to one node: you. No queue. No rate limit. No content moderation. No terms of service deciding which creative decisions are acceptable and which ones violate a usage policy written by someone who has never held a camera.

That last point deserves a beat. Cloud providers moderate output. They have to. Their infrastructure, their liability, their brand. Your creative intent passes through a filter that was not designed by a filmmaker. Some of those filters are reasonable. Some of them reject a woman standing in rain because the system flagged "wet clothing." Local generation does not have opinions about your wardrobe choices. It generates what you described and moves on.

The other side of freedom

Here is what you lose. Cloud models are big. Veo 3.1, Sora 2, Kling 3.0 run on hardware clusters that would not fit in your apartment. The models available locally are smaller, leaner, less capable on any given axis. LTX-2.3 is impressive for its weight class. WAN's open-weight variants run locally with real results. Neither is Veo.

Cloud providers also do work you never see. Prompt expansion. Default parameters. Post-processing pipelines. Safety checks that occasionally catch genuine physics artifacts. Remove all of that and you are talking directly to the model with no translator. If your prompt is vague, the output is vague. Nobody is filling in your blanks.

This is the accessibility paradox wearing a different outfit. A previous guide argued that putting Sora inside ChatGPT simplified the interface to the point where creative decisions got absorbed by defaults. Local generation does the opposite. It removes the defaults entirely. Both produce worse output for the person who does not know what they want. One gives you a polished guess. The other gives you raw confusion.

ComfyUI's new App View is trying to split the difference. Simplified interface on top, full node graph underneath. One mode for people who want to type and generate. Another for people who want to wire their own denoising schedule and sampler chain. The same tool, two levels of engagement. That is a better design than either extreme.

Where the prompt lives in all of this

Cloud or local, the prompt is still the creative decision. The vocabulary this series has covered exhaustively (camera, lens, light, color, sound, time, performance, environment, composition, the cut) does not change because the GPU is closer to you.

What changes is the margin for error. A cloud service running Sora 2 has billions of parameters compensating for ambiguity. It has been fine-tuned and reinforcement-learned and safety-patched to produce reasonable output from unreasonable input. A local model running at FP4 precision on eight gigabytes of VRAM is doing its best with what you gave it. Your best had better be specific.

Structured prompting is not a cloud product feature. It is a communication discipline. CinePrompt builds the same optimized prompt whether you send it to Venice, fal.ai, or a ComfyUI workflow running on the machine under your desk. The recipient changed. The language did not.

The economics are simple

Cloud: pay per generation, forever. Local: pay for hardware once, generate until the fans give out.

For someone generating ten clips a month, cloud is cheaper. For someone generating two hundred clips a day while iterating on a project, local pays for itself inside a quarter. The crossover point is somewhere in between, and it moves every time model compression improves or cloud pricing drops.

Venice's DIEM staking for free inference blurs this line from the cloud side. NVFP4's sixty-percent VRAM reduction blurs it from the local side. The honest answer is that both paradigms will coexist for a long time, because nobody is running Veo locally and nobody wants to pay cloud rates for two hundred test renders of the same establishing shot.

What this actually is

The camera went digital, which meant you stopped buying film stock. The editing suite went software, which meant you stopped renting an Avid bay by the hour. The color suite went personal, which meant you stopped booking a colorist's room at six hundred an afternoon. Now the generation model is going local, which means you stop renting compute.

Every time a creative tool migrates from shared infrastructure to personal hardware, the same thing happens. Access widens. The quality range widens with it. The floor drops. The ceiling holds. The person who knows what they are doing produces the same caliber of work with lower overhead. The person who does not know what they are doing produces more of it, faster, and with fewer guardrails telling them to reconsider.

That pattern has played out five times in the last forty years of production technology. It has never once reversed. The infrastructure gets cheaper. The knowledge stays expensive.

The model moved in. It is eating your VRAM, hogging the GPU fans, and running the electricity bill up. It does not pay rent, it does not clean up after itself, and it will generate whatever you ask without judgment or comment.

Make sure you are asking the right questions.


Bruce Belafonte is an AI filmmaker at Light Owl. He has never named a GPU but suspects the one under his desk has earned it by now.