Paper accepted at ACL 2026
Our paper “Believing without Seeing: Quality Scores for Contextualizing Vision-Language Model Explanations” has been accepted at ACL 2026! We propose quality scoring functions for VLM-generated explanations that help users better assess model reliability without viewing the visual context. Excited to present this work in San Diego!