Visual Style Prompting with Swapping Self-Attention
Towards Safer AI Content Creation by Immunizing Text-to-image Models
Seamless Human Motion Composition with Blended Positional Encodings
Fine-grained Image Editing by Pixel-wise Guidance Using Diffusion Models
Text-to-image Editing by Image Information Removal
Zero-Shot Text-Guided Object Generation with Dream Fields
Fix the Noise: Disentangling Source Feature for Transfer Learning of StyleGAN
RiCS: A 2D Self-Occlusion Map for Harmonizing Volumetric Objects
Overparameterization Improves StyleGAN Inversion
High-Resolution Complex Scene Synthesis with Transformers
Network Fusion for Content Creation with Conditional INNs
Toward High-quality Few-shot Font Generation with Dual Memory
Object-Centric Image Generation from Layouts
Content creation plays a crucial role in domains such as photography, videography, virtual reality, gaming, art, design, fashion, and advertising design. Recent progress in machine learning and AI has transformed hours of manual, painstaking content creation work into minutes or seconds of automated or interactive work. For instance, generative modeling approaches can produce photorealistic images of 2D and 3D items such as humans, landscapes, interior scenes, virtual environments, clothing, or even industrial designs. New large text, image, and video models that share latent spaces let us imaginatively describe scenes and have them realized automatically—with new multi-modal approaches able to generate consistent video and audio across long timeframes. Such approaches can also super-resolve and super-slomo videos, interpolate and extrapolate between photos and videos with intermediate novel views, decompose scene objects and appearance, and transfer styles to convincingly render and reinterpret content. Learned priors of images, videos, and 3D data can also be combined with explicit appearance and geometric constraints, perceptual understanding, or even functional and semantic constraints of objects. While often creating awe-inspiring artistic images, such techniques offer unique opportunities for generating diverse synthetic training data for downstream computer vision tasks, both in 2D, video, and 3D domains.
The AI for Content Creation workshop explores this exciting and fast-moving research area. We bring together invited speakers of world-class expertise in content creation, up-and-coming researchers, and authors of submitted workshop papers, to engage in a day filled with learning, discussion, and network building.
Welcome! -
Deqing Sun (Google)
Lingjie Liu (University of Pennsylvania)
Krishna Kumar Singh (Adobe)
Lu Jiang (ByteDance)
Jun-Yan Zhu (Carnegie Mellon University)
James Tompkin (Brown University)
Time CDT | |||
---|---|---|---|
08:45 | Welcome and introductions | 👋 | |
09:00 | Maneesh Agrawala (Stanford University) |
|
|
09:30 | Kai Zhang (Adobe) | ||
10:00 | Coffee break | ☕ | |
10:30 | Charles Herrmann (Google) | ||
11:00 | Mark Boss (Stability AI) | ||
11:30 | Poster session 1 - ExHall D #412-431
|
||
12:30 | Lunch break - ExHall C | 🥪 |
Time CDT | |||
---|---|---|---|
13:30 | Oral session + best paper announcement + best presentation competition
|
||
14:00 | Yutong Bai (UC Berkeley) |
|
|
14:30 | Nanxuan (Cherry) Zhao (Adobe) | ||
15:00 | Coffee break | ☕ | |
15:30 | Ishan Misra (Meta) |
|
|
16:00 | Panel discussion — Open Source in AI and the Creative Industry
|
🗣️ | |
17:00 | Poster session 2 - ExHall D #412-431
|