Collections
Discover the best community collections!
Collections including paper arxiv:2504.02542
-
AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation Framework
Paper • 2308.08155 • Published • 10 -
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence
Paper • 2401.14196 • Published • 66 -
Audio-visual Controlled Video Diffusion with Masked Selective State Spaces Modeling for Natural Talking Head Generation
Paper • 2504.02542 • Published • 51
-
One Shot, One Talk: Whole-body Talking Avatar from a Single Image
Paper • 2412.01106 • Published • 24 -
MEMO: Memory-Guided Diffusion for Expressive Talking Video Generation
Paper • 2412.04448 • Published • 10 -
IDOL: Instant Photorealistic 3D Human Creation from a Single Image
Paper • 2412.14963 • Published • 6 -
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models
Paper • 2502.01061 • Published • 222
-
Seedance 1.0: Exploring the Boundaries of Video Generation Models
Paper • 2506.09113 • Published • 102 -
Self Forcing: Bridging the Train-Test Gap in Autoregressive Video Diffusion
Paper • 2506.08009 • Published • 30 -
Seeing Voices: Generating A-Roll Video from Audio with Mirage
Paper • 2506.08279 • Published • 27 -
PolyVivid: Vivid Multi-Subject Video Generation with Cross-Modal Interaction and Enhancement
Paper • 2506.07848 • Published • 4
-
ReCamMaster: Camera-Controlled Generative Rendering from A Single Video
Paper • 2503.11647 • Published • 145 -
DreamRenderer: Taming Multi-Instance Attribute Control in Large-Scale Text-to-Image Models
Paper • 2503.12885 • Published • 43 -
TaoAvatar: Real-Time Lifelike Full-Body Talking Avatars for Augmented Reality via 3D Gaussian Splatting
Paper • 2503.17032 • Published • 27 -
Single Image Iterative Subject-driven Generation and Editing
Paper • 2503.16025 • Published • 14
-
VividFace: A Diffusion-Based Hybrid Framework for High-Fidelity Video Face Swapping
Paper • 2412.11279 • Published • 13 -
MagicFace: High-Fidelity Facial Expression Editing with Action-Unit Control
Paper • 2501.02260 • Published • 5 -
GaussianAvatar-Editor: Photorealistic Animatable Gaussian Head Avatar Editor
Paper • 2501.09978 • Published • 6 -
FantasyID: Face Knowledge Enhanced ID-Preserving Video Generation
Paper • 2502.13995 • Published • 9
-
Instruct-Imagen: Image Generation with Multi-modal Instruction
Paper • 2401.01952 • Published • 32 -
ODIN: A Single Model for 2D and 3D Perception
Paper • 2401.02416 • Published • 13 -
Bigger is not Always Better: Scaling Properties of Latent Diffusion Models
Paper • 2404.01367 • Published • 22 -
Cross-Attention Makes Inference Cumbersome in Text-to-Image Diffusion Models
Paper • 2404.02747 • Published • 13
-
Seedance 1.0: Exploring the Boundaries of Video Generation Models
Paper • 2506.09113 • Published • 102 -
Self Forcing: Bridging the Train-Test Gap in Autoregressive Video Diffusion
Paper • 2506.08009 • Published • 30 -
Seeing Voices: Generating A-Roll Video from Audio with Mirage
Paper • 2506.08279 • Published • 27 -
PolyVivid: Vivid Multi-Subject Video Generation with Cross-Modal Interaction and Enhancement
Paper • 2506.07848 • Published • 4
-
ReCamMaster: Camera-Controlled Generative Rendering from A Single Video
Paper • 2503.11647 • Published • 145 -
DreamRenderer: Taming Multi-Instance Attribute Control in Large-Scale Text-to-Image Models
Paper • 2503.12885 • Published • 43 -
TaoAvatar: Real-Time Lifelike Full-Body Talking Avatars for Augmented Reality via 3D Gaussian Splatting
Paper • 2503.17032 • Published • 27 -
Single Image Iterative Subject-driven Generation and Editing
Paper • 2503.16025 • Published • 14
-
AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation Framework
Paper • 2308.08155 • Published • 10 -
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence
Paper • 2401.14196 • Published • 66 -
Audio-visual Controlled Video Diffusion with Masked Selective State Spaces Modeling for Natural Talking Head Generation
Paper • 2504.02542 • Published • 51
-
VividFace: A Diffusion-Based Hybrid Framework for High-Fidelity Video Face Swapping
Paper • 2412.11279 • Published • 13 -
MagicFace: High-Fidelity Facial Expression Editing with Action-Unit Control
Paper • 2501.02260 • Published • 5 -
GaussianAvatar-Editor: Photorealistic Animatable Gaussian Head Avatar Editor
Paper • 2501.09978 • Published • 6 -
FantasyID: Face Knowledge Enhanced ID-Preserving Video Generation
Paper • 2502.13995 • Published • 9
-
One Shot, One Talk: Whole-body Talking Avatar from a Single Image
Paper • 2412.01106 • Published • 24 -
MEMO: Memory-Guided Diffusion for Expressive Talking Video Generation
Paper • 2412.04448 • Published • 10 -
IDOL: Instant Photorealistic 3D Human Creation from a Single Image
Paper • 2412.14963 • Published • 6 -
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models
Paper • 2502.01061 • Published • 222
-
Instruct-Imagen: Image Generation with Multi-modal Instruction
Paper • 2401.01952 • Published • 32 -
ODIN: A Single Model for 2D and 3D Perception
Paper • 2401.02416 • Published • 13 -
Bigger is not Always Better: Scaling Properties of Latent Diffusion Models
Paper • 2404.01367 • Published • 22 -
Cross-Attention Makes Inference Cumbersome in Text-to-Image Diffusion Models
Paper • 2404.02747 • Published • 13