Understanding hidden neuron activations using structured background knowledge and deductive reasoning
dc.contributor.author | Dalal, Abhilekha | |
dc.date.accessioned | 2024-11-08T20:51:38Z | |
dc.date.available | 2024-11-08T20:51:38Z | |
dc.date.graduationmonth | December | |
dc.date.issued | 2024 | |
dc.description.abstract | A central challenge in Explainable AI (XAI) is accurately interpreting hidden neuron activations in deep neural networks (DNNs). Accurate interpretations help demystify the black-box nature of deep learning models by explaining what the system internally detects as relevant in the input. While some existing methods show that hidden neuron activations can be human-interpretable, systematic and automated approaches leveraging background knowledge remain underexplored. This thesis introduces a novel model-agnostic post-hoc XAI method that integrates a Wikipedia-derived concept hierarchy of approximately 2 million classes as background knowledge and employs OWL-reasoning-based Concept Induction to generate explanations. Our approach automatically assigns meaningful class expressions to neurons in the dense layers of Convolutional Neural Networks, outperforming prior methods both quantitatively and qualitatively. In addition, we argue that understanding neuron behavior requires not only identifying what activates a neuron (recall) but also examining its precision—how it responds to other stimuli, which we define as the neuron's error margin, enhancing the granularity of neuron interpretation. To visualize these findings, we present ConceptLens, an innovative tool that visualizes neuron activations and error margins. ConceptLens offers insights into the confidence levels of neuron activations and enables an intuitive understanding of neuron behavior through visual bar charts. Together, these contributions offer a holistic approach to interpreting DNNs, advancing the explainability and transparency of AI models. | |
dc.description.advisor | Pascal Hitzler | |
dc.description.degree | Doctor of Philosophy | |
dc.description.department | Department of Computer Science | |
dc.description.level | Doctoral | |
dc.description.sponsorship | National Science Foundation grants 2119753 "RII Track-2 FEC: BioWRAP (Bioplastics With Regenerative Agricultural Properties): Spray-on bioplastics with growth synchronous decomposition and water, nutrient, and agrochemical management" 2333782 "Proto-OKN Theme 1: Safe Agricultural Products and Water Graph (SAWGraph): An OKN to Monitor and Trace PFAS and Other Contaminants in the Nation's Food and Water Systems. | |
dc.identifier.uri | https://hdl.handle.net/2097/44702 | |
dc.language.iso | en_US | |
dc.subject | Explainable AI | |
dc.subject | CNN | |
dc.subject | Neurosymbolic AI | |
dc.subject | Concept induction | |
dc.subject | Background knowledge | |
dc.title | Understanding hidden neuron activations using structured background knowledge and deductive reasoning | |
dc.type | Dissertation |