Computer Vision Meetup: Pre-trained Text2Image Diffusion Models for Zero-Shot Video Editing

Ғылым және технология

Text-to-image diffusion models demonstrate remarkable editing capabilities in the image domain, especially after Latent Diffusion Models made diffusion models more scalable. Conversely, video editing still has much room for improvement, particularly given the relative scarcity of video datasets compared to image datasets. Therefore, we will discuss whether pre-trained text-to-image diffusion models can be used for zero-shot video editing without any fine-tuning stage. Finally, we will also explore possible future work and interesting research ideas in the field.
About the Speaker
Bariscan Kurtkaya is a KUIS AI Fellow and a graduate student in the Department of Computer Science at Koc University. His research interests lie in exploring and leveraging the capabilities of generative models in the realm of 2D and 3D data, encompassing scientific observations from space telescopes.
Not a Meetup member? Sign up to attend the next event:
voxel51.com/computer-vision-a...
Recorded on June 26, 2024 at the AI, Machine Learning and Computer Vision Meetup.
#computervision #machinelearning #datascience #ai #artificialintelligence

Пікірлер

    Келесі