← Back to topics

Topic

Vision-Language Models

Vision-language models that connect text and perception.

1 papers · latest 2026-04-07

Most active fields for this topic

Jiajun Zhai, Hao Shi, Shangwei Guo et al.

cs.CVcs.MMcs.RO

E-VLA uses event cameras—normally used in robotics—to let robots see and act in near-total darkness or blur, where normal cameras fail. This enables real-world robotic systems to operate reliably in challenging environments like smoke-filled rooms or fast-moving scenes.

© 2026 A2A.pub — AI to Action. From papers to practice, daily.
Summaries are AI-assistedPrivacyTerms