[{"data":1,"prerenderedAt":502},["ShallowReactive",2],{"content-/zh/advanced-tutorial/wan2.2":3},{"id":4,"title":5,"body":6,"description":494,"extension":495,"meta":496,"navigation":497,"path":498,"seo":499,"stem":500,"__hash__":501},"content/zh/advanced-tutorial/wan2.2.md","万相2.2（Wan2.2）完整教程：阿里通义实验室最强开源视频模型",{"type":7,"value":8,"toc":470},"minimark",[9,13,26,29,34,41,45,91,93,97,100,174,176,180,183,187,194,216,220,262,264,268,271,274,349,351,355,358,361,411,413,417,421,432,436,462,464,467],[10,11,5],"h1",{"id":12},"万相22wan22完整教程阿里通义实验室最强开源视频模型",[14,15,16,17,21,22,25],"p",{},"本篇指南全面拆解由阿里通义实验室推出的最新一代开源视频生成模型 ",[18,19,20],"strong",{},"Wan2.2","。无论你是专业动态设计师还是 AI 视频爱好者，把 Wan2.2 接入 ",[18,23,24],{},"ComfyUI"," 后，都能实现电影级 AI 视频的精细化控制。",[27,28],"hr",{},[30,31,33],"h2",{"id":32},"一wan22-是什么架构上的一次革命","一、Wan2.2 是什么？架构上的一次革命",[14,35,36,37,40],{},"Wan2.2 于 2026 年初发布，不只是 Wan2.1 的小版本迭代，而是首次在视频扩散模型中引入了 ",[18,38,39],{},"MoE（混合专家）架构","。",[42,43,44],"h3",{"id":44},"核心技术亮点",[46,47,48,69,79,85],"ul",{},[49,50,51,54,55],"li",{},[18,52,53],{},"MoE 混合专家架构","\n传统大模型是“稠密型”，每次计算都会激活全部参数；\nWan2.2 采用双专家机制：",[46,56,57,63],{},[49,58,59,62],{},[18,60,61],{},"高噪声专家","：负责初期画面构图与运动规划",[49,64,65,68],{},[18,66,67],{},"低噪声专家","：负责后期精细纹理与细节渲染",[49,70,71,74,75,78],{},[18,72,73],{},"高效能设计","\n模型总参数量 27B，但",[18,76,77],{},"实时激活仅 14B","，\n用小得多的显存占用，跑出顶级画质。",[49,80,81,84],{},[18,82,83],{},"电影级美学","\n训练数据量比上一代多 80% 以上，专门针对光影、对比度、专业运镜做了标注，成片更像实拍影视。",[49,86,87,90],{},[18,88,89],{},"原生高分辨率","\n支持直接生成 720p、1280p 分辨率，24fps 流畅视频。",[27,92],{},[30,94,96],{"id":95},"二模型版本对比14b-版-vs-5b-版","二、模型版本对比：14B 版 vs 5B 版",[14,98,99],{},"根据你的显卡配置选择即可：",[101,102,103,123],"table",{},[104,105,106],"thead",{},[107,108,109,114,117,120],"tr",{},[110,111,113],"th",{"align":112},"left","模型",[110,115,116],{"align":112},"激活参数量",[110,118,119],{"align":112},"建议显存",[110,121,122],{"align":112},"适用场景",[124,125,126,143,158],"tbody",{},[107,127,128,134,137,140],{},[129,130,131],"td",{"align":112},[18,132,133],{},"Wan2.2-T2V-A14B",[129,135,136],{"align":112},"14B",[129,138,139],{"align":112},"24GB 以上（RTX 3090/4090）",[129,141,142],{"align":112},"高端电影级文生视频",[107,144,145,150,152,155],{},[129,146,147],{"align":112},[18,148,149],{},"Wan2.2-I2V-A14B",[129,151,136],{"align":112},[129,153,154],{"align":112},"24GB 以上",[129,156,157],{"align":112},"专业图生视频，时序一致性极强",[107,159,160,165,168,171],{},[129,161,162],{"align":112},[18,163,164],{},"Wan2.2-TI2V-5B",[129,166,167],{"align":112},"5B",[129,169,170],{"align":112},"10GB–12GB",[129,172,173],{"align":112},"消费级显卡快速迭代测试",[27,175],{},[30,177,179],{"id":178},"三comfyui-安装教程","三、ComfyUI 安装教程",[14,181,182],{},"运行 Wan2.2 需要最新版 ComfyUI + 专属自定义节点。",[42,184,186],{"id":185},"步骤1安装自定义节点","步骤1：安装自定义节点",[14,188,189,190,193],{},"打开 ",[18,191,192],{},"ComfyUI Manager","，搜索安装：",[195,196,197,204,210],"ol",{},[49,198,199,203],{},[200,201,202],"code",{},"ComfyUI-WanVideoWrapper","（作者 Kijai）：目前 Wan2.2 最稳定的封装节点",[49,205,206,209],{},[200,207,208],{},"ComfyUI-VideoHelperSuite","：加载图片、导出 MP4/GIF 必备",[49,211,212,215],{},[200,213,214],{},"ComfyUI-KJNodes","：提供专用遮罩与噪声工具",[42,217,219],{"id":218},"步骤2模型文件放置","步骤2：模型文件放置",[46,221,222,236,245],{},[49,223,224,227,228,231,232,235],{},[18,225,226],{},"主扩散模型","：放到 ",[200,229,230],{},"ComfyUI/models/checkpoints/","\n（部分节点版本也可放在 ",[200,233,234],{},"models/diffusion_models/","）",[49,237,238,227,241,244],{},[18,239,240],{},"VAE",[200,242,243],{},"ComfyUI/models/vae/","\n必须用 Wan2.2 专用 VAE，才能正确支持 16×16×4 压缩格式",[49,246,247,250,251,254,255,258,259],{},[18,248,249],{},"文本编码器","：一般需要 ",[18,252,253],{},"T5-v1.1-xxl"," 和 ",[18,256,257],{},"UMT5","，放到 ",[200,260,261],{},"models/clip/",[27,263],{},[30,265,267],{"id":266},"四文生视频t2v工作流教程","四、文生视频（T2V）工作流教程",[14,269,270],{},"从零生成视频，需要规范的提示词 + 合理采样参数。",[42,272,273],{"id":273},"节点搭建",[195,275,276,286,307,319],{},[49,277,278,281,282,285],{},[18,279,280],{},"WanVideo Loader","：选择 ",[200,283,284],{},"Wan2.2-T2V-14B"," 模型",[49,287,288,291,292],{},[18,289,290],{},"Empty Wan Latent","：设置分辨率与帧数\n",[46,293,294,300],{},[49,295,296,297],{},"14B 模型推荐：",[18,298,299],{},"1280×720",[49,301,302,303,306],{},"帧数：",[18,304,305],{},"81 帧","（24fps 下约 5 秒）",[49,308,309,312,313],{},[18,310,311],{},"CLIP Text Encode","：Wan2.2 更懂自然语句，别堆关键词\n优质提示词示例：\n",[314,315,316],"blockquote",{},[14,317,318],{},"暴雨中的赛博朋克都市，电影级跟拍镜头，霓虹灯光倒映在水洼中，超写实，8K，高对比度",[49,320,321,324,325],{},[18,322,323],{},"高级 KSampler"," 设置\n",[46,326,327,330,333,343],{},[49,328,329],{},"步数：30–50",[49,331,332],{},"CFG：5.0–7.0（Wan2.2 对 CFG 敏感，别设太高）",[49,334,335,336,339,340],{},"采样器：",[200,337,338],{},"uni_pc"," 或 ",[200,341,342],{},"euler",[49,344,345,346],{},"调度器：",[200,347,348],{},"simple",[27,350],{},[30,352,354],{"id":353},"五图生视频i2v工作流教程","五、图生视频（I2V）工作流教程",[14,356,357],{},"Wan2.2 的图生视频是 2026 年公认的“行业标杆”，\n能让 AI 静图动起来，且时序一致性极强。",[42,359,360],{"id":360},"分步搭建",[195,362,363,369,377,383,398],{},[49,364,365,368],{},[18,366,367],{},"Load Image","：导入高清底图（如 Midjourney、Flux 生成的图）",[49,370,371,281,374],{},[18,372,373],{},"WanVideo I2V Loader",[200,375,376],{},"Wan2.2-I2V-14B",[49,378,379,382],{},[18,380,381],{},"Image-to-Latent","：将图片接入 WanVideo I2V Encoder，转为模型可理解的潜变量",[49,384,385,388,389,392,393],{},[18,386,387],{},"提示词","：只描述",[18,390,391],{},"动作","即可\n示例：\n",[314,394,395],{},[14,396,397],{},"人物转头看向镜头并微笑，风吹动发丝",[49,399,400,403],{},[18,401,402],{},"Motion Bucket（运动强度）",[46,404,405,408],{},[49,406,407],{},"高值（80+）：动作幅度大、动态强烈",[49,409,410],{},"低值（30–50）：适合人像微动态、柔和镜头",[27,412],{},[30,414,416],{"id":415},"六高级优化技巧","六、高级优化技巧",[42,418,420],{"id":419},"使用-lightx2v-v2-lora-加速","使用 Lightx2v V2 LoRA 加速",[14,422,423,424,427,428,431],{},"如果生成速度太慢，可以加载 ",[18,425,426],{},"Lightx2v V2"," 蒸馏 LoRA，\n能把采样步数从 ",[18,429,430],{},"40 步直接降到 8–12 步","，画质几乎不下降，\n对 RTX 3080、4070 这类单显卡用户极其友好。",[42,433,435],{"id":434},"显存爆了这样优化","显存爆了？这样优化",[46,437,438,448,455],{},[49,439,440,441,339,444,447],{},"在加载器中开启 ",[200,442,443],{},"fp8",[200,445,446],{},"bf16"," 精度",[49,449,450,451,454],{},"使用 ",[18,452,453],{},"VAE Tile Encode"," 分块编码，避免一次性加载超大帧",[49,456,457,458,461],{},"先降到 ",[18,459,460],{},"832×480"," 做预览版，确认效果再拉高分辨率",[27,463],{},[30,465,466],{"id":466},"总结",[14,468,469],{},"Wan2.2 让高端视频生成真正走向普及。\n借助 MoE 架构 + ComfyUI 可视化工作流，你可以做出媲美商业闭源工具的电影级短片，无论是个人创作还是商用项目都完全够用。",{"title":471,"searchDepth":472,"depth":472,"links":473},"",2,[474,478,479,483,486,489,493],{"id":32,"depth":472,"text":33,"children":475},[476],{"id":44,"depth":477,"text":44},3,{"id":95,"depth":472,"text":96},{"id":178,"depth":472,"text":179,"children":480},[481,482],{"id":185,"depth":477,"text":186},{"id":218,"depth":477,"text":219},{"id":266,"depth":472,"text":267,"children":484},[485],{"id":273,"depth":477,"text":273},{"id":353,"depth":472,"text":354,"children":487},[488],{"id":360,"depth":477,"text":360},{"id":415,"depth":472,"text":416,"children":490},[491,492],{"id":419,"depth":477,"text":420},{"id":434,"depth":477,"text":435},{"id":466,"depth":472,"text":466},"本篇指南全面拆解由阿里通义实验室推出的最新一代开源视频生成模型 Wan2.2。无论你是专业动态设计师还是 AI 视频爱好者，把 Wan2.2 接入 ComfyUI 后，都能实现电影级 AI 视频的精细化控制。","md",{},true,"/zh/advanced-tutorial/wan2.2",{"title":5,"description":494},"zh/advanced-tutorial/wan2.2","SthYSTYzKfht7AMKULk_1VIoWTuUF_d_WaA_RoPEI1s",1773986044745]