Topic
Diffusion Models
Diffusion-based generation for images, video, and multimodal outputs.
2 papers · latest 2026-04-08
Most active fields for this topic
Hiba Dahmani, Nathan Piasco, Moussab Bennehar et al.
SEM-ROVER enables scalable, geometrically coherent 3D driving scene generation via semantic voxel-guided diffusion—enabling realistic, large-scale simulation for autonomous driving systems without view limitations.
Hyunsoo Cha, Wonjung Woo, Byungjun Kim et al.
Vanast eliminates the need for separate try-on and animation steps by doing both in one go, reducing distortions and identity drift. This means you can generate realistic, coherent videos of people wearing new clothes from just one image—useful for e-commerce and virtual fashion without complex pipelines.