All the papers listed in this project come from my usual reading. If you have found some new and interesting papers, I would appreciate it if you let me know!!!
Are SAE features from the Base Model still meaningful to LLaVA? (Dec. 6, 2024)
Bridging the VLM and mech interp communities for multimodal interpretability (Oct. 28, 2024)
Laying the Foundations for Vision and Multimodal Mechanistic Interpretability & Open Problems (May. 14, 2024)
-
LMM Concept Explainability A Concept-Based Explainability Framework for Large Multimodal Models (Nov. 30, 2024)
-
Attention Lens Devils in Middle Layers of Large Vision-Language Models: Interpreting, Detecting and Mitigating Object Hallucinations via Attention Lens (Nov. 23, 2024)
-
SAE 4 LMM Large Multi-modal Models Can Interpret Features in Large Multi-modal Models (Nov. 22, 2024)
-
LLaVA in VQA Understanding Multimodal LLMs: the Mechanistic Interpretability of Llava in Visual Question Answering (Nov. 17, 2024)
-
What Do VLMs NOTICE? What Do VLMs NOTICE? A Mechanistic Interpretability Pipeline for Gaussian-Noise-free Text-Image Corruption and Evaluation (Oct. 18, 2024)
-
LLaVA-Intrepret Towards Interpreting Visual Information Processing in Vision-Language Models (Oct. 09, 2024)
-
LVLM-Intrepret LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models (June. 24, 2024)
-
Multi-Modal Neurons Finding and Editing Multi-Modal Neurons in Pre-Trained Transformers (Jun. 11, 2024)
-
Information Storage and Transfer Understanding Information Storage and Transfer in Multi-modal Large Language Models (June. 06, 2024)
-
Causal Tracing Tool 4 BLIP Towards Vision-Language Mechanistic Interpretability: A Causal Tracing Tool for BLIP (Aug. 27, 2023)