Wan 2.2: Alibaba’s Tongyi Lab Releases the World’s First Open-Source MoE Video Generation Model
In the dynamic realm of AI video generation, Wan 2.2 stands out as the world's first open-source Mixture-of-Experts (MoE) architecture video generation model, unveiled by Alibaba's Tongyi Lab on July 28, 2025. Often referred to as wan2.2 or simply wan video, this multimodal powerhouse excels in text to video (T2V), image to video (I2V), and hybrid tasks, offering cinematic-level control over lighting, composition, color grading, and complex motions like hip-hop dancing or street parkour. Fully open-sourced under an Apache 2.0 license on GitHub, Hugging Face, and ModelScope, Wan 2.2 models support up to 720p resolutions at 24fps, running efficiently on consumer-grade GPUs like the RTX 4090, making it ideal for digital art, advertising, film previsualization, and game development.
The wan 2.2 models family features three variants: Wan2.2-T2V-A14B (14B parameters for superior T2V with MoE-driven layout and detail refinement), Wan2.2-I2V-A14B (14B for stable I2V synthesis reducing artifacts in stylized scenes), and Wan2.2-TI2V-5B (5B hybrid for fast 720p generation via 16×16×4 compression). This upgrade surpasses Wan 2.1 in motion fidelity—achieving reliable camera movements like pan left/right, dolly in/out, and orbital arcs—and benchmarks like Wan-Bench 2.0, where it tops competitors in semantics and aesthetics. With WanBox for all-in-one creation and editing, Wan 2.2 embodies "All in Wan, Create Anything," inviting global innovation in open video AI.