15.01.24, 18:00 Uhr
These models can be easily trained to recognize inferences made by black-box AI systems and explain them via natural language texts that are easily comprehensible to users. The quality of these explanations might also be improved through reinforcement learning mechanisms that leverage users’ feedback. However, the use of LLMs as XAI tools introduces unexpected risks and presents us with new challenges. Unlike conventional XAI techniques, LLMs themselves are equally, if not more, opaque than the ML models they should explain. This raises an additional “meta” black-box problem: how can one places trust in an explanation provided by an LLM-based XAI tool if one does not understand how this explanation has been generated?
In his lecture, Alberto Termine will explore more in-depth the challenges and risks raised by the use of LLMs for XAI purposes. He will particularly focus on the phenomenon of hallucination, i.e., the tendency of LLMs to produce explanations that seem reliable but, in reality, include a plethora of false and potentially misleading information. He will contend that hallucination severely undermines the reliability of LLMs as XAI tools, and that addressing this issue constitutes the primary challenge in the development of safe, fair, and trustworthy LLM-based XAI systems.