ServicesAI governance
What is explainability in AI?
Quick links
Art deco aesthetic, minimalist image of a desk lamp shining light onto a transparent black box. Inside the box, intricate circuitry is visible, illuminated by the soft glow of the lamp. The design emphasizes geometric shapes, clean lines, and symmetry typical of art deco style, symbolizing transparency and revealing inner workings. Futuristic and clean, without any text or symbols—By Talbot West

What is explainability in AI?

By Jacob Andra / Published October 10, 2024 
Last Updated: October 10, 2024

Executive summary:

Most AI systems function as "black boxes," with opaque decision-making processes. This lack of transparency poses significant risks, especially in sectors where failure could be damaging or even catastrophic.

AI explainability seeks to illuminate the decision-making processes of AI systems so that humans can better understand, steer, and govern them. Benefits of explainability include the following:

  • Increased trust in AI systems
  • Improved regulatory compliance
  • Enhanced ability to detect and correct biases
  • Easier debugging and improvement of AI models
  • Better AI governance and reduced legal liability

Explainable AI is an emerging set of architectures, practices, and techniques that help organizations understand and interpret AI decisions, practice good governance, and comply with regulations. Methods range from feature attribution techniques to visualization tools and modular AI architectures such as cognitive hive AI (CHAI).

Ready to explore how explainable AI can enhance your organization's AI capabilities? Contact Talbot West for a free consultation on implementing explainable AI tailored to your specific needs and industry requirements.

BOOK YOUR FREE CONSULTATION

Artificial intelligence (AI) is becoming increasingly interwoven into government, business operations, and other critical systems. Most AI systems function as a black box, in which the decisions of the AI (and its processes to reach those decisions) are opaque.

AI has the potential to wreak havoc, especially in industries such as finance, health, or defense. On the milder end of the spectrum, it might decline a transaction, make a faulty diagnosis, or flag the wrong target. More catastrophic (and somewhat futuristic yet plausible) scenarios include shutting down power grids, launching weapons, crashing financial markets, and any number of other scary outcomes.

To circumvent AI-driven catastrophe and error, explainability seeks to illuminate the processes by which AI makes its decisions. With more transparency, humans can better steer AI away from bad outcomes and toward good outcomes.

Explainable AI is an emerging set of architectures, practices, and techniques that help organizations understand and interpret AI decisions, practice good governance, and comply with regulations.

Main takeaways
AI decision-making is opaque.
A lack of AI accountability stifles implementation in sensitive industries.
Uninterpretable AI represents a potential hazard.
Explainability is a set of techniques to understand how AI makes decisions.
Cognitive hive AI (CHAI) represents a new paradigm for AI explainability

The black box problem is bad for business

AI models (particularly deep learning systems) take in data, process it through complex layers of mathematical operations, and produce outputs without providing insight into their decision-making process. Opacity exists due to the intricate nature of these models, which involve millions (or billions) of parameters and non-linear relationships that defy simple explanations.

The black box problem stems from the characteristics that make modern AI so powerful. Deep neural networks excel at finding subtle patterns in vast amounts of data, but their internal workings become obscure as they grow more sophisticated. Even the engineers who design these systems often can't explain why a particular input leads to a specific output.

This lack of transparency slows AI adoption and implementation for the following reasons:

  • Trust deficit: Users, whether customers or internal stakeholders, are reluctant to rely on decisions they can't understand or verify.
  • Regulatory compliance: Many industries require explainable decision-making processes, particularly in sectors such as finance and healthcare.
  • Bias detection: Without visibility into the model's reasoning, it's challenging to identify and correct for biases in the training data or algorithm.
  • Debugging and improvement: When a model performs poorly, uninterpretability makes it difficult to pinpoint and address the root causes.
  • Governance concerns: As AI systems make increasingly impactful decisions, the inability to scrutinize their reasoning is an ethical problem.
  • Legal liability: In cases where AI decisions lead to negative outcomes, organizations that can’t explain their perspective face significant legal exposure.

What are popular explainable AI methods?

The following methods attempt to make AI systems and machine learning models understandable. The goal of these techniques is to generate outputs that explain AI decisions to human users.

Feature attribution methods

Feature attribution techniques identify which input features contributed most significantly to a particular prediction. SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations) are two popular feature attribution methods.

SHAP assigns each feature an importance value for a particular prediction. It's based on game theory and provides consistent, locally accurate post-hoc explanations.

While SHAP can work with many models, it’s best for tree-based models. For deep learning models, it can be computationally expensive.

LIME creates a simpler, interpretable model that approximates the behavior of the complex model around a specific instance. This local model can then explain which features were most important for that particular model prediction.

LIME is model-agnostic, but it’s mostly used for explaining predictions of complex neural networks.

LIME-SUP (LIME with supervised perturbations)

This is an extension of LIME that uses domain knowledge to create more realistic perturbations, leading to more accurate and relevant explanations.

LIME-SUP has niche applications that handle complex models in domains with well-understood feature relationships.

Example-based explanations

This approach finds similar instances in the training data to explain individual predictions. For instance, if an AI flags a transaction as fraudulent, it might show similar fraudulent transactions from the past to generate post-hoc explanations.

Grad-CAM (gradient-weighted class activation mapping)

Primarily used on image classification models, Grad-CAM generates heat maps highlighting the regions of an image that were most important for the model's decision.

Grad-CAM is designed for convolutional neural networks and doesn't apply to other models.

Partial dependence plots (PDP)

PDPs show how a feature affects predictions on average, helping to understand the relationship between input features and the model's output.

PDPs work with many models but can be computationally intensive for large datasets or complex models.

Decision trees as surrogate models

For very complex models, a simpler decision tree can be trained to mimic the behavior of the black box model, providing an approximation of its decision process.

Decision trees are useful for approximating complex models, but the fidelity of the approximation fluctuates depending on the complexity of the original model.

Counterfactual explanations

These explain what changes would be necessary to alter the model's prediction. For example, "If a transaction was $50 instead of $100, it wouldn't have been flagged as suspicious."

Counterfactual explanations are model-agnostic but see heavy use in decision-making processes where understanding alternative outcomes is needed (e.g., loan approvals, fraud detection).

While these explainability techniques are quite technical, they are a necessary component of successfully implementing AI-based systems. And forward-thinking organizations aren’t shying away from illuminating opaque models to seize the full power of AI.

Cognitive hive AI for explainability

One exciting breakthrough in LLM explainability comes from the use of multiple LLMs (along with other types of AI or other forms of knowledge management) working together in a modular architecture. This approach, which we at Talbot West are calling cognitive hive AI (CHAI), offers a significant leap forward in our ability to understand and trust AI decision-making processes.

Here's how CHAI increases explainability:

  • Modular architecture: AI stacks leverage a network of smaller, specialized language models that collaborate to process information. This modular structure makes it easier to examine how individual components contribute to the final output.
  • Reasoning transparency: When dealing with complex problems, AI stacks can provide insights into their thought process, helping users understand how the AI arrived at its conclusions.
  • Explainability-first design: From the ground up, AI stacks are more transparent and interpretable than traditional large language models. You can query the constituent AI components to see how the overall system arrived at a conclusion.

Beyond enhanced explainability, CHAI has some other advantages over monolithic LLMs:

  • Resource-friendly operation: You can design a lightweight AI stack to run on everyday devices, such as personal computers or mobile phones.
  • Independence from cloud services: Unlike many AI models that rely on remote servers, you can deploy AI stacks offline, such as an air-gapped deployment for sensitive defense applications.
  • Budget-friendly implementation: You can deploy CHAI customized to your specific parameters, rather than paying extra for functionality you don’t need.
  • Streamlined performance: A well-designed CHAI can outperform traditional LLMs in efficiency, potentially delivering quicker results while using fewer resources.
  • Data protection: With local processing capabilities, CHAI offers an extra layer of security and privacy over cloud-dependent models.

Explainable AI in business

Let’s look at some specific AI implementations in which explainability plays an important role.

Art deco aesthetic, minimalist image showing the upper torso and head of a robot making a prediction using a glowing crystal ball. The crystal ball has stylized buildings inside, representing industries like finance, healthcare, and construction. Light rays or abstract energy waves emanate from the crystal ball, symbolizing the predictive power of the robot. The robot is futuristic with clean geometric lines, emphasizing symmetry and clarity, without any text or symbols—By Talbot West

Financial services (credit scoring)

In finance, banks use machine learning models to assess loan applications. These models analyze hundreds of variables, from credit history to social media activity. Regulations such as the Equal Credit Opportunity Act require lenders to provide specific reasons for adverse actions. Here, SHAP values break down the model's decision.

A denied applicant might learn that their debt-to-income ratio contributed 40% to the negative decision, while their recent job change contributed 30%. This explanation satisfies regulatory requirements and helps applicants understand how to improve their chances in the future.

Healthcare (medical imaging diagnostics)

In healthcare, hospitals employ convolutional neural networks (CNNs) to detect anomalies in medical images such as X-rays or MRIs. To increase the transparency of CNNs, radiologists use Grad-CAM to generate heatmaps highlighting the areas of an image that most influenced the model's diagnosis.

In a chest X-ray flagged for potential pneumonia, Grad-CAM might highlight areas of lung opacity that the model focused on. This visual explanation allows radiologists to quickly validate the AI's findings and focus their attention on relevant areas. Example-based explanations also provide similar cases from the training data, helping less experienced radiologists understand the model's decision in the context of historical cases.

Manufacturing (predictive maintenance)

In manufacturing, factory operators use ensemble methods such as random forests to predict equipment failures based on sensor data. These models are challenging to interpret, making it difficult for maintenance teams to properly plan interventions. To solve this, manufacturers implement partial dependence plots (PDPs) to understand how different sensor readings impact failure predictions.

A PDP might reveal that vibration levels above a certain threshold dramatically increase failure probability. Computer scientists also apply SHAP values to help rank the importance of different sensors for each prediction. This allows maintenance teams to focus on the most critical indicators and understand the interplay between various factors. Explainable AI helps factories move beyond simple "failure imminent" warnings to actionable insights that guide precise, timely maintenance interventions.

Customer service (chatbot response generation)

E-commerce companies use large language models to power customer service chatbots. These models generate human-like responses to customer queries but sometimes produce inappropriate or incorrect answers. To mitigate this risk, companies implement a two-step explainable AI approach combined with human-in-the-loop features.

Human agents use post-hoc explanation methods such as LIME to identify which parts of the customer's input most influenced the chatbot's response. This helps debug unusual responses and improve the model's context understanding. Then, they employ counterfactual explanations to generate alternative responses based on slight changes in the input.

Human supervisors can quickly review and select the most appropriate response in ambiguous situations. Explainable AI lets companies maintain the efficiency of AI-driven customer service, exercise quality control, and build user trust.

Best practices for RAG implementation

Talbot West is here to guide you through every step of the implementation process. Here are the aspects we emphasize for our clients:

  • Start small. Begin with a manageable subset of your data to launch a pilot project and demonstrate proof of concept. After we test and refine your RAG system, it’s time to scale up.
  • Prioritize data quality. Your RAG system is only as good as the data you feed it. That’s why we focus heavily on document preprocessing so that your RAG is fed the most delicious, easily-digestible information possible.
  • Monitor and iterate. Test your RAG system and monitor its performance. Iterate as needed with prompt engineering and custom instructions.
  • Be responsible. We'll help you navigate ethical concerns and implement RAG responsibly with a solid governance framework.

At Talbot West, we don't just implement RAG—we partner with you to create a solution tailored to your unique business needs. Our end-to-end support covers everything from initial setup to ongoing optimization.

How to implement explainable AI

Explainable AI (XAI) will help you get the most out of your organization's AI strategy. Here's how you can user confidence and accountability through explainable AI:

  • Start with the end in mind: Before diving into XAI techniques, clearly define what you need to explain and to whom. A data scientist might need detailed feature importance, while a customer may just want to know why their loan was denied.
  • Choose the right tools: Several platforms offer XAI capabilities out of the box. Google Cloud's Vertex AI, for example, provides built-in explainability features for its machine learning models. For open-source solutions, libraries such as SHAP and LIME are effective. For the ultimate in explainability, explore a custom cognitive hive AI deployment (talk to Talbot West for feasibility and details).
  • Integrate XAI into your AI development lifecycle: Don't treat explainability as an afterthought. Incorporate it from the model design phase. This might mean choosing more interpretable models or planning for explainability techniques from the start.
  • Train your team: Your data scientists and ML engineers need to understand XAI techniques and their importance. They should be able to generate explanations and correctly interpret them.
  • Create explanation interfaces: Develop user-friendly dashboards or reports that present AI explanations in an accessible format. This might include visualizations such as feature importance charts or example-based explanations.
  • Establish an XAI review process: Set up a system to regularly review model explanations. This can catch potential biases or errors before they impact business decisions.
  • Bridge the gap between technical and business teams: Foster collaboration between data scientists and business stakeholders. Regular meetings to discuss model explanations can help AI align with business goals and ethics.
  • Document everything: Maintain clear records of your models, their explanations, and decisions made based on these explanations. Audits and court cases might depend on your documentation. Documentation also helps improve your model.
  • Test with real users: Regularly gather feedback from the end-users of your AI systems. Are the explanations helpful? Understandable? Accurate? Use this feedback to refine your approach.

By making explainability a core part of your AI strategy, you'll build more trustworthy, effective, and valuable AI systems. The best way to successfully add explainability to your AI is to work with a partner that understands the process.

Work with Talbot West

At Talbot West, we guide companies through the AI implementation process, including AI explainability. We help with:

  • AI feasibility: We run a feasibility study to determine how useful AI systems will be for your business.
  • AI proof of concept: We’ll do a limited pilot project to demonstrate that the system works and can generate a positive ROI.
  • AI strategy: We develop a cohesive AI strategy that includes explainability and human-in-the-loop features.
  • AI tool selection: We help you select the right tools and interfaces for your AI and interpret its decisions.
  • AI implementation: We help you set up the framework, preprocess data, and calibrate your XAI tools or CHAI configuration.
  • AI ethics and governance: We verify that your AI tools and usage patterns are legal, ethical, and explainable.

Our mission is to blend your team's expertise and AI's capabilities, empowering your workforce and unleashing your organization’s potential.

Ready to explore XAI tools for your organization?

AI FAQ

An example of explainable artificial intelligence is a credit risk assessment model for small business loans that uses a gradient-boosting algorithm.

The model provides a risk score and uses SHAP values to show how each factor (e.g., cash flow, credit history, time in business) contributes to the final score. It also offers what-if scenarios, showing how improving specific factors changes the risk assessment. This process helps loan officers and applicants understand the decision process and potential paths to approval.

Explainable artificial intelligence methods in natural language processing include:

Attention visualization, which highlights words or phrases the model focuses on when making predictions.
Layer-wise relevance propagation, which traces the contribution of each input word through the neural network layers.
Saliency maps identify which parts of a sentence most influenced the model's output.
For text classification tasks, LIME generates local explanations by altering the input text and observing changes in the model's predictions.

Monolithic “black box” LLMs are, by nature, not very explainable. Their explainability can be enhanced with some of the methods we discussed earlier, but there will always be a limit. At Talbot West, we believe that true explainability will come about through a modular, “hive” AI architecture that we’re calling “cognitive hive AI,” or CHAI.

CHAI spreads internal functions across modules, which allows a much higher level of transparency.

Explainable models are needed, especially in high-stakes domains such as healthcare, finance, and criminal justice. Explainable AI builds trust, enables regulatory compliance, and facilitates debugging and improving model performance. It helps detect and mitigate biases, enhancing fairness in AI-driven decisions.

About the author

Jacob Andra is the founder of Talbot West and a co-founder of The Institute for Cognitive Hive AI, a not-for-profit organization dedicated to promoting Cognitive Hive AI (CHAI) as a superior architecture to monolithic AI models. Jacob serves on the board of 47G, a Utah-based public-private aerospace and defense consortium. He spends his time pushing the limits of what AI can accomplish, especially in high-stakes use cases. Jacob also writes and publishes extensively on the intersection of AI, enterprise, economics, and policy, covering topics such as explainability, responsible AI, gray zone warfare, and more.
Jacob Andra

Industry insights

We stay up to speed in the world of AI so you don’t have to.
View All

Subscribe to our newsletter

Cutting-edge insights from in-the-trenches AI practicioners
Subscription Form

About us

Talbot West bridges the gap between AI developers and the average executive who's swamped by the rapidity of change. You don't need to be up to speed with RAG, know how to write an AI corporate governance framework, or be able to explain transformer architecture. That's what Talbot West is for. 

magnifiercrosschevron-downchevron-leftchevron-rightarrow-right linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram