Back to feed
UX
Trend

Explainable AI in Chat Interfaces

Nielsen Norman Group

Article Overview

The Nielsen Norman Group article, "Explainable AI in Chat Interfaces," critically examines the efficacy of explanation text in AI chatbots, arguing that current practices frequently fall short of helping users genuinely understand AI outputs. As AI chat interfaces become increasingly prevalent, users rely heavily on their responses for decision-making, yet the underlying AI systems remain "black boxes" without proper explanations. The core thesis is that effective explanations are crucial for users to form accurate mental models (internal representations of how something works), prevent the spread of misinformation, and make informed decisions about trusting AI outputs. However, the article points out that explanations provided by current LLMs (Large Language Models - AI models trained on vast amounts of text data) are often inaccurate, hidden, or confusing, exacerbating issues like overtrust.

A significant challenge highlighted is the inherent technical complexity of modern AI models, which are so intricate that even AI engineers struggle to trace the exact reasoning behind every output. This technical limitation means that true, transparent explainability is an ongoing industry conversation. Despite this, AI chatbots often present their outputs with high confidence, leading users to place undeserved trust in potentially inaccurate or hallucinated (AI generating plausible but false information) answers. The article specifically focuses on textual explanations like source citations, step-by-step walkthroughs, and disclaimers, analyzing their intended benefits versus their actual risks in interfaces such as Claude, ChatGPT, Copilot, and Gemini.

Source citations, intended for verification, are frequently problematic. They can be hallucinated, point to non-existent URLs, or link to irrelevant/unreliable content, giving users a false sense of reliability. Research on AI information-seeking behaviors reveals that users rarely click these citation links, trusting the mere presence of a citation. For instance, DeepSeek was observed providing a seemingly credible link that led to a 404 error. To mitigate this, UX teams are advised to set realistic expectations, style citations prominently, place them directly next to the supported claim, link to relevant parts of sources, and use meaningful labels (e.g., publication name) instead of generic ones like "Source." Copilot and Perplexity are cited as examples of better practices, displaying citations as clickable chips or inline tags.

Step-by-step reasoning, while appearing to offer transparency, is often a post-hoc rationalization rather than a faithful representation of the model's actual computation. Research indicates these explanations can omit influencing factors or adjust to justify incorrect answers, leading to plausible but untrue narratives. This presents a dilemma for designers: while such explanations can make a product feel approachable, they risk promoting false trust. The article concludes that given their unreliability, designers should avoid step-by-step explanations that imply certainty, instead prioritizing sources and clear disclaimers about limitations until the technology for true explainability matures. Disclaimers, though theoretically useful, are often ignored by users who skim content, further complicating the communication of AI limitations.

Impact on Design Practice

This article profoundly impacts UX/UI designers by shifting the focus from simply *providing* information to *designing for critical engagement* with AI outputs. Designers can no longer assume that adding a "source" link or a "disclaimer" will suffice; they must actively design interfaces that encourage skepticism and verification. This means moving beyond passive information display to creating interactive elements that prompt users to question, explore, and validate AI-generated content.

For instance, instead of burying citations at the bottom of a response, designers should treat them like critical evidence in a courtroom. They need to be prominently displayed, contextually linked to specific claims, and styled to invite interaction, much like an interactive footnote. This requires a fundamental shift in how designers approach information architecture and visual hierarchy within AI chat interfaces. Furthermore, the revelation that step-by-step explanations are often rationalizations, not true insights, challenges designers to resist the urge to create seemingly transparent but ultimately misleading features. It pushes them to prioritize honesty about AI's current limitations over a false sense of explainability, potentially influencing the adoption of more cautious design patterns for complex AI tasks.

Ultimately, designers are now tasked with becoming "trust architects," not just "information presenters." Their daily work will involve more rigorous user testing around how explanations are perceived and acted upon, and a deeper collaboration with AI engineers to understand the true capabilities and limitations of the models they are designing for. This ensures that the user's mental model of the AI system is accurate, preventing overreliance and promoting responsible interaction.

Design AI explanations to actively encourage critical verification, not just passive consumption, by making sources prominent, contextual, and actionable, and by honestly communicating model limitations.

How to Apply This

The article offers concrete strategies for improving how source citations are presented in AI chat interfaces to encourage user verification and build more accurate mental models.

1

Set Realistic Expectations: Clearly state within the interface that users must verify sources as they may be inaccurate or fabricated.

2

Prominently Display Citations: Avoid burying sources; make them highly visible and distinct from the main response to encourage investigation.

3

Contextually Link Sources: Place citations directly alongside the specific claims or sentences they support, making in-context verification easier.

4

Link to Relevant Content: When possible, link directly to the precise section of an article that supports the claim, reducing user effort in finding information.

5

Use Meaningful Labels: Replace generic "Source" labels with descriptive titles like the publication name or article title, informing users what they are clicking.

AI Chatbot Examples

The article mentions several prominent AI chat interfaces as examples of current practices in explanation text:

* **Claude:** An AI assistant developed by Anthropic. * **ChatGPT:** A large language model chatbot developed by OpenAI. * **Copilot:** An AI assistant integrated into various Microsoft products. * **Gemini:** A multimodal AI model developed by Google. * **DeepSeek:** An AI model, specifically mentioned for a problematic citation example.

Industry Context

This article fits squarely within the broader industry conversation around Explainable AI (XAI) and responsible AI design. As AI models become more powerful and integrated into everyday tools, the challenge of making their outputs understandable and trustworthy is paramount. This piece highlights the gap between the technical capabilities of LLMs and the user's need for transparency, underscoring the critical role of UX design in bridging this divide. It reflects a growing industry awareness that simply deploying AI is insufficient; designing for user comprehension, managing expectations, and mitigating risks like misinformation and overtrust are equally important for AI's ethical and effective adoption. The discussion on "hallucinations" and the limitations of current "explainability" mechanisms are central themes in contemporary AI development and UX discourse.