UniReal: Universal Image Generation and Editing via Learning Real-world Dynamics
Abstract
We introduce UniReal, a unified framework designed to address various image generation and editing tasks. Existing solutions often vary by tasks, yet share fundamental principles: preserving consistency between inputs and outputs while capturing visual variations. Inspired by recent video generation models that effectively balance consistency and variation across frames, we propose a unifying approach that treats image-level tasks as discontinuous video generation. Specifically, we treat varying numbers of input and output images as frames, enabling seamless support for tasks such as image generation, editing, customization, composition, etc. Although designed for image-level tasks, we leverage videos as a scalable source for universal supervision. UniReal learns world dynamics from large-scale videos, demonstrating advanced capability in handling shadows, reflections, pose variation, and object interaction, while also exhibiting emergent capability for novel applications.
Community
Hi, thanks for your work. Is there a planned date for code release?
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Diffusion Self-Distillation for Zero-Shot Customized Image Generation (2024)
- VIVID-10M: A Dataset and Baseline for Versatile and Interactive Video Local Editing (2024)
- One Diffusion to Generate Them All (2024)
- AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea (2024)
- FoundHand: Large-Scale Domain-Specific Learning for Controllable Hand Image Generation (2024)
- OmniEdit: Building Image Editing Generalist Models Through Specialist Supervision (2024)
- HyperGAN-CLIP: A Unified Framework for Domain Adaptation, Image Synthesis and Manipulation (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper