Why Does Explainability Matter in News Analytic Systems? Proposing Explainable Analytic Journalism
Source of Publication
As the use of algorithms has emerged in journalism, analytic/algorithmic journalism (AJ) has seen rapid development in major news organizations. Despite this surging trend, little is known about the role and the effects of explainability on the process by which people perceive and make sense of trust in an algorithm-driven AI system. While AJ has greatly benefited from increasingly sophisticated algorithm technologies, AJ suffers from a lack of transparency and understandability for readers. We identify explainability as a heuristic cue of an algorithm and conceptualizes it in relation to trust by testing how it affects user emotion with AJ. Our experiments show that the addition of interpretable explanations leads to enhanced trust in the context of AJ and readers' trust hinges upon the perceived normative values that are used to assess algorithmic qualities. Explanations of why certain news articles are recommended give users emotional assurance and affirmation. Mediation analyses show that explanatory cues play a mediating role between trust and performance expectancy. The results have implications for the inclusion of explanatory cues in AJ, which help to increase credibility and help users to assess AJ value.
Social and Behavioral Sciences
analytic journalism, explainable algorithmic journalism, Explainable journalism, explanatory cues, interpretability, news personalization, understandability
Shin, Donghee, "Why Does Explainability Matter in News Analytic Systems? Proposing Explainable Analytic Journalism" (2021). All Works. 4287.
Indexed in Scopus