Fantasy AIGC Family is an open-source initiative exploring Human-centric AI, World Modeling, and Human-World Interaction, aiming to bridge perception, understanding, and generation in the real and digital worlds.
- 🎉 Nov, 2025: FantasyTalking2 and FantasyHSI are accepted by AAAI 2026.
- 👋 Aug, 2025: We release the inference code and model weights of FantasyPortrait.
- 🎉 Jul, 2025: FantasyTalking is accepted by ACM MM 2025.
- 👋 Apr, 2025: We release the inference code and model weights of FantasyTalking, FantasyID.
The first Wan-based high-fidelity audio-driven avatar system that synchronizes facial expressions, lip motion, and body gestures in dynamic scenes through dual-stage audio-visual alignment and controllable motion modulation.
🗣️ FantasyTalking2: Timestep-Layer Adaptive Preference Optimization for Audio-Driven Portrait Animation
A novel Timestep-Layer Adaptive Multi-Expert Preference Optimization (TLPO) method enhances the quality of audio-driven avatar in three dimensions: lip-sync, motion naturalness, and visual quality.
🗿 FantasyHSI: Video-Generation-Centric 4D Human Synthesis In Any Scene through A Graph-based Multi-Agent Framework
A graph-based multi-agent framework that grounds video generation within 3D world dynamics, enabling digital humans to perceive, plan, and act autonomously, thus serving as the technical bridge that links human modeling to world modeling through unified perception–action reasoning.
🤡 FantasyPortrait: Enhancing Multi-Character Portrait Animation with Expression-Augmented Diffusion Transformers
A novel expression-driven video-generation method that pairs emotion-enhanced learning with masked cross-attention, enabling the creation of high-quality, richly expressive animations for both single and multi-portrait scenarios.
A unified world model integrating video priors and geometric grounding for synthesizing explorable and geometrically consistent 3D scenes.
A tuning-free text-to-video model that leverages 3D facial priors, multi-view augmentation, and layer-aware guidance injection to deliver dynamic, identity-preserving video generation.
- Giving Back to the Community: In our daily work, we benefit immensely from the resources, expertise, and support of the open source community, and we aim to give back by making our own projects open source.
- Attracting More Contributors: By open sourcing our code, we invite developers worldwide to collaborate—making our models smarter, our engineering more robust, and extending benefits to even more users.
- Building an Open Ecosystem: We believe that open source brings together diverse expertise to create a collaborative innovation platform—driving technological progress, industry growth, and broader societal impact.