Explore projects
- 
Updated
 - 
Updated
 - 
Gradio demo showcasing VL-SHAP. You can generate visually informed explanations of textual outputs for VL models
Updated - 
Updated
 - 
[Frontiers in AI Journal] Implementation of the paper "Interpreting Vision and Language Generative Models with Semantic Visual Priors"
Updated - 
Targeted semantic multimodal input ablation. Official implementation of the ablation method introduced in the paper: "What Vision-Language Models 'See' when they See Scenes"
Updated - 
[INLG2023] The High-Level (HL) dataset is a Vision and Language (V&L) resource aligning object-centric descriptions from COCO with high-level descriptions crowdsourced along 3 axes: scene, action, rationale.
Updated - 
Updated
 - 
Updated
 - 
Updated
 - 
Updated
 - 
Updated