The video landscape changed hard in the first quarter of 2026. Kling released version 3.0. Veo 3.1 added 4K and scene extension. Runway shipped Gen-4.5. Seedance 2.0 landed with multi-shot consistency. Sora announced its own closure.

Seven models matter right now. Each one is genuinely good at something different. This is the short version of which to pick, when, and why.

The quick version

If you want the answer before the explanation:

StrengthModelWhy
Closest replacement for SoraKling 3.0Multi-shot, 4K, native audio, 15s clips. The cleanest one-to-one move.
Dialogue and lipsyncVeo 3.1Audio generated alongside the image, not layered on after. Google’s Gemini app and Flow.
Narrative, directed, film-lookRunway Gen-4.5Mature web interface. Best control tools for directed, multi-step edits.
Physical realism and human motionSeedance 2.0Strong on physics and body mechanics. Western access via Flora only.
Character close-ups and micro-expressionsMiniMax Hailuo 2.3Six-second cap by design. Short, dense, beautifully animated.
Long-form generation via extensionLuma Ray3.14Up to 18 seconds per clip. Three times cheaper than Ray3.
Leaving the fieldSora 2OpenAI closes the app on April 26, 2026. Migration guide here.

Now the detail.


Kling 3.0 (Kuaishou)

Kuaishou released Kling 3.0 on February 5, 2026. Reports from post-Sora migration coverage suggest it is one of the models working creators are reaching for most — and in our testing it is the cleanest one-to-one replacement for Sora’s multi-shot work.

What Kling does best: Multi-shot composed sequences with integrated audio. Fifteen seconds per generation, native 4K at 60 frames a second, and audio — dialogue, lipsync, sound design — generated alongside the video rather than added after the fact. Kling 3.0 is also unusually good at persisting a world across shots, which is the thing that made Sora feel more like film than flipbook.

Where it falls short: Fifteen seconds is not sixty. For longer sequences you still need to plan a shot list and assemble in a timeline. The audio is generated, which means it is approximate — you will still want to do a post-production pass. And Kling’s pricing is credit-based rather than a clean monthly-tier list, which takes a little time to read if you are coming in cold.

Pricing: Credit-based via klingai.com. No single flat per-clip price — consumption depends on resolution, duration, and variant.

Access: klingai.com (web), Flora Fauna (which carries Kling 3.0 Pro and Standard alongside earlier Kling versions), and through API partners.


Veo 3.1 (Google)

Google released Veo 3.1 in October 2025 and the cost-reduced Veo 3.1 Lite on the Gemini API in March 2026. The version you use in the Gemini app or in Flow is the full Veo 3.1.

What Veo does best: Audio. Specifically, the fact that audio is generated at the same time as the image — synchronised dialogue, sound effects, and ambient soundscapes, all native to the generation. No other model does this as cleanly. If your scene involves a character saying something, Veo is the best choice by a meaningful margin. Full Veo workflow guide here.

Where it falls short: Clips are four, six, or eight seconds per generation. You can extend a scene iteratively to build longer sequences, but you cannot ask for a single thirty-second shot. Character consistency across a full scene is good but not perfect — expect to regenerate a few times to land the exact look you want.

Resolution: Up to 4K, added in January 2026. Vertical 9:16 for mobile-first formats is also supported.

Pricing: Free tier via Google Vids (10 generations a month on any personal Google account). Google AI Pro at $19.99 a month. Google AI Ultra at $249.99 a month. API pricing from $0.05 per second on Veo 3.1 Lite up to $0.40 per second on Veo 3.1 Standard.

Access: The Gemini app, Flow (flow.google), Google Vids, the Gemini API, and Flora Fauna (which carries Veo 3.1, Fast, Lite, and Frames variants).


Runway Gen-4.5

Runway released Gen-4.5 on December 1, 2025. It is the model most closely associated with directed, narrative work — the kind of film you sit down to plan.

What Runway does best: Control. Gen-4.5 has the most mature set of what Runway calls “control modes” — image-to-video (animate a still), keyframes (pin specific moments the generation must hit), video-to-video (feed it an existing clip and restyle it), and a growing set of directed-generation tools. If your project is a short film or a music video or anything where the visual language is the whole point, Runway is the model that most looks like it was directed.

Where it falls short: Native audio was added to Gen-4.5 in December 2025 — dialogue, sound effects, and ambient — but Veo 3.1 and Kling 3.0 still lead on lip-sync precision. Clips are shorter than Kling — typically in the range of a few seconds to around ten per generation. Pricing is per-credit on top of the monthly plan, which can add up fast on production work.

Pricing: Monthly plans start around $15, with higher tiers increasing credit allocations. Annual billing reduces the monthly rate.

Access: runwayml.com (web), API, and Flora Fauna. Runway Aleph — Runway’s newer editing-focused tool — is also carried on Flora as a separate model.


Seedance 2.0 (ByteDance)

Seedance 2.0 released in February 2026. On paper it is exceptional. In practice, Western access is the complicated part.

What Seedance does best: Multi-shot consistency and physically accurate human motion. Seedance 2.0 sits near the top of current public video benchmarks — it is widely regarded as one of the leading models for human realism and body mechanics.

Where it falls short: Availability. ByteDance paused the global launch of Seedance 2.0 in mid-March 2026. CapCut integration rolled out only to Brazil, Indonesia, Malaysia, Mexico, the Philippines, Thailand, and Vietnam. The Dreamina app is restricted. Direct API access is limited to fal.ai.

For UK, European, and US creators, the practical route right now is through Flora Fauna, which carries both Seedance 2.0 and Seedance 2.0 Fast. Flora has become the cleanest way to use this model if you are outside the live regions.

Access: Flora Fauna (recommended for Western creators), fal.ai API, Dreamina and CapCut in select markets.


MiniMax Hailuo 2.3

Hailuo is the outlier on this list. Six-second clips, by design. The MiniMax team has been explicit that the cap is intentional.

What Hailuo does best: Short, dense, emotionally precise moments. Character close-ups, micro-expressions, small gestures. If you need the exact beat of a laugh, a glance, a breath — the kind of shot that is six seconds because six seconds is all it needs to be — Hailuo 2.3 is often the best choice. Its Media Agent feature lets you hand it an image, a clip, or a sound to shape what it makes.

Where it falls short: Six seconds. That is not a weakness so much as a philosophical choice. You cannot use Hailuo for a thirty-second establishing shot. It is built for the single beat, not the sequence.

Pricing: Matched to the previous Hailuo 02 generation. Credit-based via hailuoai.video.

Access: hailuoai.video, Flora Fauna (carries Hailuo 2.3 Pro), and partner APIs.


Luma Ray3.14

Luma released Ray3.14 on January 26, 2026. It is the upgrade to Luma’s Ray 3 line, and notable for being significantly cheaper — Luma’s own announcement cites three times lower cost than Ray 3.

What Luma does best: Longer generations and flexible editing. Up to eighteen seconds per clip, with image-to-video using first-and-last-frame references (you give it the opening and closing images, and it generates the motion between). Native 1080p. If you want a single long take rather than a multi-shot assembly, Luma is worth trying.

Where it falls short: Audio is not confirmed as a native feature on Ray3.14 — you will add sound in post. And Ray3.14 is not on Flora Fauna as of April 2026. Flora currently carries Luma Ray 2 and Ray 2 Flash only — so if you use Flora as your main workspace, you will not have the newest Luma inside it.

Pricing: Credit-based via lumalabs.ai with a free tier for experimentation.

Access: lumalabs.ai (web), Dream Machine app, partner APIs.


Sora 2 — the one leaving

OpenAI announced on March 24, 2026 that the Sora app will shut down on April 26, 2026, with the API following on September 24, 2026. There is no successor product. Sora will continue as internal research on world models, but the thing you use to make videos is going away.

If you have Sora work or were planning to use it this month, read the migration guide for export instructions and the cleanest moves to Kling, Veo, or Runway.


How to choose

The honest answer: most of these models can do most things. The difference is in the details.

Audio is the first question. If your scene involves dialogue, lipsync, or a specific soundscape — Veo 3.1. No other model handles audio as cleanly. Kling 3.0 is a strong second with native audio in its generations, but Veo is still the specialist.

Length is the second. If you need a single long take, Luma Ray3.14 (up to 18s) or Kling 3.0 (15s) give you the most headroom per generation. Veo’s 8s cap is fine for scene-work with extension, less fine if you want one continuous shot. Hailuo 2.3 is deliberately short.

Motion and physics is the third. If your shot is a person running, a body falling, a cloth catching the wind, a ball bouncing — these are the things that early video models got wrong most often. Seedance 2.0 currently leads most of the public motion benchmarks. Kling 3.0 is the strong generalist runner-up. Veo holds its own. Luma is worth trying for flowing, atmospheric motion.

Control is the fourth. If your workflow is image-to-video, keyframes, or video-to-video — Runway Gen-4.5 has the most mature tools. If you want multi-shot consistency without hand-holding — Kling 3.0 or Seedance 2.0.

Consistency across shots is the fifth. Generating one beautiful clip is easy. Generating six that feel like they belong to the same scene is the hardest thing any of these models try to do. Kling 3.0 and Seedance 2.0 are the current leaders. Veo 3.1 is improving. No model solves this perfectly — for a long-form project you will still select, regenerate, and sometimes composite across models.

Access is the sixth. If you want a single subscription that covers every model on this list, Flora Fauna is the cleanest option. It carries Kling 3.0, Veo 3.1, Runway Gen-4.5, Seedance 2.0, Hailuo 2.3, Luma Ray 2, and the legacy Sora 2 Pro entry (which will stop working after April 26) — all from one node-based canvas. For Seedance 2.0 specifically, it is currently the only practical route for Western creators.

The practical approach: Most working video creators do not use a single model. They use two or three for what each one is best at. Kling for the wide establishing shots, Veo for the dialogue scene, Hailuo for the reaction close-up. It is a less satisfying answer than “just pick the winner” — but it is where almost every working creator lands.

What the benchmarks say

The most-cited public video benchmarks right now are Artificial Analysis’s Video Arena and Tsinghua’s VBench. They measure different things and do not always agree. Both currently place Seedance 2.0 and Kling 3.0 near the top, with Veo 3.1 and Runway Gen-4.5 close behind. As with any benchmark, treat the rankings as a starting point, not a verdict — the right model for your brief is the one that produces the shot you can actually use.


What about open source?

The open-source video ecosystem is real and growing. Alibaba’s Wan 2.2 is the most interesting project for most creators — released July 2025 under a clean Apache 2.0 licence, which means you can use it commercially without a lawyer. Weights are on Hugging Face. A single RTX 4090 with 24GB of VRAM is enough for the smaller TI2V-5B model.

Tencent’s HunyuanVideo 1.5 (released November 2025) is technically impressive, but the licence is restrictive: it excludes the UK, EU, and South Korea entirely. If you are reading this from any of those places, Hunyuan is not a legal option for commercial work, even though the weights are public.

The trade-off with open source is the same as it always is: more control, more setup, and a quality gap against the leading proprietary models that is real but narrowing. If you are running your own pipeline on serious hardware and you are outside the excluded territories, Wan is the one to start with. If you are not, the seven models above will cover most briefs.


Art & Algorithms publishes guides, tutorials, and prompt packs at the intersection of art and code. Subscribe for the full archive.