What Is Wan 2.2 and Should You Use It for AI Video?

Share article

Most AI video models can show you a cat on a skateboard. Wan 2.2 lets you decide how it’s lit, how the camera moves, and what kind of lens you’re pretending to shoot it with. In this post, we’ll break down what it does differently… and whether it’s actually worth your time.

Tip:

Wan 2.2 is now available in getimg.ai's Video Generator! Try it now in Text to Video and Image to Video modes.

All You Need to Know About Wan 2.2

Wan 2.2 is a Text to Video and Image to Video model released by Alibaba’s Tongyi Lab in July 2025. It builds on the foundation laid by Wan 2.1 but pushes things further, especially when it comes to motion quality, prompt responsiveness, and overall visual richness.

The biggest leap? Training data. Wan 2.2 wasn’t just fed more clips, it was fed better ones. The dataset grew by over 65% in still images and more than 80% in video, but it’s the labeling that matters. Each clip came annotated with precise visual metadata: lighting conditions, color grading, camera movement, framing, texture, and more.

Tracking shot of a man sprinting through a dusty attic, running straight into a tall mirror. As he hits the glass, [match cut] he bursts through on the other side, now running across a vast desert — same speed, same motion, without breaking stride.

This is why Wan 2.2 doesn’t just generate a man walking through a city at night. Because it knows the difference, it gives you a slow pan across wet pavement, moody sidelighting, and a soft lens flare from passing headlights.

A two-expert system

Under the hood, Wan 2.2 introduces a more flexible way of thinking. Instead of relying on one massive model to do everything, it uses a Mixture of Experts (MoE) system: two specialized components take turns handling different parts of the generation process.

The first expert sets the stage, laying down structure, movement, spatial depth. The second swoops in to refine the details, sharpen textures, and clean up the noise. It’s like having a cinematographer work alongside a compositor, with each playing to their strengths.

That split also makes the model more efficient. While the total architecture clocks in at 27 billion parameters, only a portion is active at any given time. So you get a large-model feel, without the GPU meltdown.

Slow tilt down from a cloudy sky to reveal a post-apocalyptic city in ruins, faint ash drifting in the air, sunlight breaking through gaps in destroyed buildings, camera continuing downward to frame a lone figure standing in the street

Camera-aware prompting

Wan 2.2 isn’t just built to respond to prompts, it’s built to take direction. Thanks to a built-in system called VACE 2.0 (short for Video Animation Control Engine), it understands and responds to motion language in a way most open models can’t.

That means prompts like:

  • “Aerial orbit around a snowy cabin at dusk”
  • “Handheld tracking shot through a neon-lit alley”
  • “Zoom-in on a character’s face, warm backlight, shallow focus”

…actually work, more often than not. You can prompt for lighting style, camera type, lens effect, even mood, and get results that reflect that input. It’s not perfect (nothing in this space is), but it’s impressively good at hitting the tone you ask for.

A boxer in a gritty gym throws slow punches toward the camera, which performs a subtle arc around him from left to right, catching dust drifting through a warm shaft of light above the ring.


Can you run it yourself?

Technically, yes. But unless you like debugging your own driver errors, you probably shouldn’t.

Wan 2.2 includes multiple variants, including a compact 5B hybrid version (TI2V‑5B) that can run on a single RTX 4090. But even then, you’ll need to manage compression settings, memory limits, model configs, and enough Python tools to fill a weekend.

So yes, it’s open source. But that doesn’t make it user-friendly. Which is exactly why we’ve added it to getimg.ai’s Video Generator. Now you can use Wan 2.2 directly in your browser, with no extra setup.

How to make an AI video with Wan 2.2?

All you need to do is:

  1. Open the Video Generator and choose the Text to Video or Image to Video mode.
  2. Select Wan 2.2 from the model picker.
  3. Write a prompt and (in Image to Video mode) upload a starting frame.
  4. Generate!

A bride sprints barefoot across a windswept cliffside road toward the groom, veil trailing behind her, as ocean waves crash below. The groom lifts her and spins her in his arms. A white dress flutters in the wind, cutting sharply against the stormy blue sky. Smooth side-tracking camera follows the action from medium distance, with dramatic backlight and wide cinematic framing.

Bottom line

Wan 2.2 doesn’t try to be everything, but what it does, it does surprisingly well, especially if you’re looking for more control over how your video looks, not just what it shows.

Now that it’s available in getimg.ai’s Video Generator, joining other powerful models like Veo 3 and Hailuo 02, you don’t need to choose between “powerful” and “easy to use.” We can’t wait to see what you come up with!




Get started with getimg.ai

Create an account and start creating AI content for free. Work smarter, not harder.

Like creating with AI?

Earn getimg.ai credits for generating and sharing beautiful content.

Join Program

Have questions or feedback?

We're here to help.

Contact us