Visual Classification as Linear Combination of Words

Explainability is a longstanding challenge in deep learning, especially in high-stakes domains like healthcare. Common explainability methods highlight image regions that drive an AI model’s decision. Humans, however, heavily rely on language to convey explanations of not only “where” but “what”. Additionally, most explainability approaches focus on explaining individual AI predictions, rather than describing the features used by an AI model in general.

Data and Resources

Cite this as

Shobhit Agarwal, Yevgeniy R. Semenov, William Lotter (2024). Dataset: Visual Classification as Linear Combination of Words. https://doi.org/10.57702/x6l6hv1y

DOI retrieved: December 3, 2024

Additional Info

Field Value
Created December 3, 2024
Last update December 3, 2024
Defined In https://doi.org/10.48550/arXiv.2311.10933
Author Shobhit Agarwal
More Authors
Yevgeniy R. Semenov
William Lotter
Homepage https://github.com/lotterlab/task_word_explainability