This paper investigates the sensitivity analysis of word importance in generative AI models, specifically using the GPT model and attention weights. It introduces a novel approach that ranks word importance by integrating Kullback-Leibler divergence with sensitivity analysis, aiming to enhance interpretability and accountability of transformer-based models. Through the use of the Cornell Movie-Dialogs corpus, the study seeks to provide a deeper understanding of model performance and bridge the gap between complex AI systems and explainable AI (XAI) methodologies.
Related topics: