Explainable AI Through Combination of Deep Tensor and Knowledge Graph

Explainable AI means humans can understand why conclusion was reached or how machines make decisions or specific judgments.

Explainability of AI (XAI) is fundamental for people to trust, manage and use AI models. People will not use AI systems if they don’t get explainability.
AI is called to present both its judgments and the reasons for them at the same time.

Our Partner Fujitsu Laboratories of America has developed the very first explainable AI, called Deep Tensor, capable of showing the reasons behind AI-generated findings and to make them explainable, allowing human experts to validate the truth of AI produced results and gain new insights.

Deep Tensor with Knowledge Graph is already in use in several mission-critical businesses: health management (health changes are detected from employee activity data for workstyle transformation) and investment decisions in finance (corporate growth is predicted based on KPIs).

For example, in genomic medicine, Fujitsu Laboratories of America has built an AI system by training Deep Tensor by inputting 180,000 pieces of disease genetic mutation data and associating more than 10 billion pieces of knowledge from 17 million medical articles, etc. in Knowledge Graph. Medical specialists simply review the inference logic flow, which greatly cut the time from analysis to report submission (2weeks to 1 day).

A portion of the data used in validating the effectiveness of this technology in the field of genomic medicine was an outcome of joint development with Kyoto University in relation to “Construction of Clinical Genome Knowledge Base to Promote Precision Medicine” in the “Program for an Integrated Database of Clinical and Genome Information” under the Japan Agency for Medical Research and Development (AMED).

Fujitsu’s new deep learning technology connects the proprietary AI technology Deep Tensor, which performs machine learning on graph-structured data, with graph-structured knowledge bases i.e knowledge graph.

Fig. 1 Deep Tensor + Knowledge Graph -> Reasons and basis of AI judgments

Connecting inferences derived by Deep Tensor to Knowledge Graph enables to understand the reasons behind AI-generated findings and to make them explainable.

The paper below explains how Deep Tensor works in terms of application cases in network intrusion detection and genomic medicine.

Fig 2. Deep Tensor enables Explainable AI

The full version is available at this link



Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
LARUS Business Automation

LARUS Business Automation


We are data engineers and data scientists team passionate about data and top quality insight-driven software solutions.