Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Jan 20;12(1):1040.
doi: 10.1038/s41598-021-04590-0.

A review of some techniques for inclusion of domain-knowledge into deep neural networks

Affiliations

A review of some techniques for inclusion of domain-knowledge into deep neural networks

Tirtharaj Dash et al. Sci Rep. .

Abstract

We present a survey of ways in which existing scientific knowledge are included when constructing models with neural networks. The inclusion of domain-knowledge is of special interest not just to constructing scientific assistants, but also, many other areas that involve understanding data using human-machine collaboration. In many such instances, machine-based model construction may benefit significantly from being provided with human-knowledge of the domain encoded in a sufficiently precise form. This paper examines the inclusion of domain-knowledge by means of changes to: the input, the loss-function, and the architecture of deep networks. The categorisation is for ease of exposition: in practice we expect a combination of such changes will be employed. In each category, we describe techniques that have been shown to yield significant changes in the performance of deep neural networks.

PubMed Disclaimer

Conflict of interest statement

The authors declare no competing interests.

Figures

Figure 1
Figure 1
An example of AI for Science. The human-in-the-loop is a biologist. The biologist conducts experiments in a biological system, obtains experimental observations. The biologist then extracts data that can be used to construct machine learning model(s). Additionally, the machine learning system has access to domain knowledge that can be obtained from the biologist. The machine learning system then conveys its explanations to the biologist.
Figure 2
Figure 2
The plots from showing gains in predictive accuracy of (a) multilayer perceptron (MLP), and (b) graph neural network (GNN) with the inclusion of domain-knowledge. The domain knowledge inclusion method in (a) is a simple technique known as ‘propositionalisation’ ; and, the method in (b) is a general technique of incorporating domain-knowledge using bottom-graph construction. The results shown are over 70 datasets. No importance to be given to the line joining two points; this is done for visualisation purpose only.
Figure 3
Figure 3
Informal descriptions of (a) logical; and (b) numerical constraints.
Figure 4
Figure 4
Construction of a deep model M from data (D) using a learner (L). We use π to denote the structure (organisation of various layers, their interconnections etc.) and θ to denote the parameters (synaptic weights) of the deep network. L denotes the loss function (for example, cross-entropy loss in case of classification).
Figure 5
Figure 5
Some implications of using domain-knowledge for commonly-used deep network architectures. Although attention-mechanism has also been used recently in many deep network architectures, we mention it only for RNNs and transformers as it is more prominently being used for sequence learning.
Figure 6
Figure 6
Introducing background knowledge into deep network by transforming data. T is a transformation block that takes input data D, background knowledge (BK) and outputs transformed data D that is then used to construct a deep model using a learner L.
Figure 7
Figure 7
Introducing background knowledge into deep network by transforming the loss function L. T block takes an input loss L and outputs a new loss function L by transforming (augmenting or modifying) L based on background knowledge (BK). The learner L then constructs a deep model using the original data D and the new loss function L.
Figure 8
Figure 8
Introducing background knowledge into deep network by transforming the model (structure and parameter). In (a), the transformation block T takes a input structure of a model π and outputs a transformed structure π based on background knowledge (BK). In (b), the transformation block T takes a set of parameters θ of a model and outputs a transformed set of parameters π based on background knowledge (BK).

References

    1. Stevens, R. et al. Ai for science. Tech. Rep., Argonne National Lab.(ANL), Argonne, IL (United States) (2020).
    1. Kitano H. Artificial intelligence to win the nobel prize and beyond: Creating the engine for scientific discovery. AI Mag. 2016;37:39–49.
    1. Lipton, Z. C. The mythos of model interpretability. arXiv preprint arXiv:1606.03490 (2016).
    1. Arrieta, A. B. et al. Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai. arXiv preprint arXiv:1910.10045 (2019).
    1. Dash, T., Srinivasan, A. & Vig, L. Incorporating symbolic domain knowledge into graph neural networks. Mach. Learn. 1–28 (2021).