← Back to topics
Topic
Vision-Language Models
Vision-language models that connect text and perception.
1 papers · latest 2026-04-07
Most active fields for this topic
Jiajun Zhai, Hao Shi, Shangwei Guo et al.
cs.CVcs.MMcs.RO
E-VLA uses event cameras—normally used in robotics—to let robots see and act in near-total darkness or blur, where normal cameras fail. This enables real-world robotic systems to operate reliably in challenging environments like smoke-filled rooms or fast-moving scenes.