TY - CHAP
T1 - EXPLAINABLE AI AND AWE
T2 - Balancing Tensions between Transparency and Predictive Accuracy
AU - Boulanger, David
AU - Kumar, Vivekanandan Suresh
N1 - Publisher Copyright:
© 2024 selection and editorial matter, Mark D. Shermis and Joshua Wilson; individual chapters, the contributors.
PY - 2024/1/1
Y1 - 2024/1/1
N2 - Automated writing evaluation (AWE) is a complex research topic having a long way to go before becoming a mature and trustworthy technology. Research focuses on reporting the performance of AWE systems by measuring their predictive accuracy without analyzing what the algorithm actually learns. This lack of transparency is partly due to the high complexity and dimensionality of AWE, and the usage of deep learning to cope with them just exacerbated the problem. Interestingly, the ascent of explainable artificial intelligence (XAI) allows to look retrospectively at the way AWE is developed. However, XAI is still in its infancy facing many limitations. Consequently, it is crucial to understand these limitations and how to correctly interpret black-box scoring models. This chapter investigates the extent to which XAI can help determine the most generalizable model and its explanations help fix learned fallacies. The chapter presents several applications and cautions about XAI in the AWE field, drawing conclusions that (1) the explanation models of the random forests had much higher descriptive accuracy than those of the gradient-boosted trees, two inherently opaque architectures, and (2) the simplest AWE models produced the most stable, consistent, and interpretable explanations.
AB - Automated writing evaluation (AWE) is a complex research topic having a long way to go before becoming a mature and trustworthy technology. Research focuses on reporting the performance of AWE systems by measuring their predictive accuracy without analyzing what the algorithm actually learns. This lack of transparency is partly due to the high complexity and dimensionality of AWE, and the usage of deep learning to cope with them just exacerbated the problem. Interestingly, the ascent of explainable artificial intelligence (XAI) allows to look retrospectively at the way AWE is developed. However, XAI is still in its infancy facing many limitations. Consequently, it is crucial to understand these limitations and how to correctly interpret black-box scoring models. This chapter investigates the extent to which XAI can help determine the most generalizable model and its explanations help fix learned fallacies. The chapter presents several applications and cautions about XAI in the AWE field, drawing conclusions that (1) the explanation models of the random forests had much higher descriptive accuracy than those of the gradient-boosted trees, two inherently opaque architectures, and (2) the simplest AWE models produced the most stable, consistent, and interpretable explanations.
UR - http://www.scopus.com/inward/record.url?scp=85195770569&partnerID=8YFLogxK
U2 - 10.4324/9781003397618-26
DO - 10.4324/9781003397618-26
M3 - Chapter
AN - SCOPUS:85195770569
SN - 9781032502564
SP - 445
EP - 468
BT - The Routledge International Handbook of Automated Essay Evaluation
ER -