-
Viscept - 看见概念
- Shanghai
- http://www.huxiuhan.com
Starred repositories
Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding
1 min voice data can also be used to train a good TTS model! (few shot voice cloning)
Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions
Makes the prompts for generating anime images more detailed by upsampling Danbooru tags.
We introduce a novel approach for parameter generation, named neural network parameter diffusion (p-diff), which employs a standard latent diffusion model to synthesize a new set of parameters
☁️ VRAM for SDXL, AnimateDiff, and upscalers. Run your workflows on the cloud, from your local ComfyUI
[CVPR 2024] X-Adapter: Adding Universal Compatibility of Plugins for Upgraded Diffusion Model
[SIGGRAPH ASIA 2024 TCS] AnimateLCM: Computation-Efficient Personalized Style Video Generation without Personalized Video Data
[ICML 2024] Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (RPG)
Building a quick conversation-based search demo with Lepton AI.
InstantID: Zero-shot Identity-Preserving Generation in Seconds 🔥
This are a series of ComfyUI workflows that work together to create and repurpose animation
Instant voice cloning by MIT and MyShell.
Unleash endless possibilities with ComfyUI and Stable Diffusion, committed to crafting refined AI-Gen tools and cultivating a vibrant community for both developers and users.
StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation
An image/video/workflow browser and manager for ComfyUI.
Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation
[ICCV 2023] Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
AnimateDiff for AUTOMATIC1111 Stable Diffusion WebUI