Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2017 Apr 26;3(4):283-293.
doi: 10.1021/acscentsci.6b00367. Epub 2017 Apr 3.

Low Data Drug Discovery with One-Shot Learning

Affiliations

Low Data Drug Discovery with One-Shot Learning

Han Altae-Tran et al. ACS Cent Sci. .

Abstract

Recent advances in machine learning have made significant contributions to drug discovery. Deep neural networks in particular have been demonstrated to provide significant boosts in predictive power when inferring the properties and activities of small-molecule compounds (Ma, J. et al. J. Chem. Inf.

Model: 2015, 55, 263-274). However, the applicability of these techniques has been limited by the requirement for large amounts of training data. In this work, we demonstrate how one-shot learning can be used to significantly lower the amounts of data required to make meaningful predictions in drug discovery applications. We introduce a new architecture, the iterative refinement long short-term memory, that, when combined with graph convolutional neural networks, significantly improves learning of meaningful distance metrics over small-molecules. We open source all models introduced in this work as part of DeepChem, an open-source framework for deep-learning in drug discovery (Ramsundar, B. deepchem.io. https://github.com/deepchem/deepchem, 2016).

PubMed Disclaimer

Conflict of interest statement

The authors declare no competing financial interest.

Figures

Figure 1
Figure 1
Schematic of Network Architecture for one-shot learning in drug discovery.
Figure 2
Figure 2
Pictorial depiction of iterative refinement of embeddings. Inputs/outputs are two-dimensional for illustrative purposes, with q1 and q2 forming the coordinate axes. Red and blue points depict positive/negative samples (for illustrative purposes only). The original embedding g′(S) is shown as squares. The expected features r are shown as empty circles.
Figure 3
Figure 3
Graphical representation of the major graph operations described in this paper. For each of the operations, the nodes being operated on are shown in blue, with unchanged nodes shown in light blue. For graph convolution and graph pool, the operation is shown for a single node, v; however, these operations are performed on all nodes v in the graph simultaneously.

References

    1. Ma J.; Sheridan R. P.; Liaw A.; Dahl G. E.; Svetnik V. Deep neural nets as a method for quantitative structure-activity relationships. J. Chem. Inf. Model. 2015, 55, 263–274. 10.1021/ci500747n. - DOI - PubMed
    1. Ramsundar B. deepchem.io. https://github.com/deepchem/deepchem, 2016.
    1. Waring M. J.; Arrowsmith J.; Leach A. R.; Leeson P. D.; Mandrell S.; Owen R. M.; Pairaudeau G.; Pennie W. D.; Pickett S. D.; Wang J.; Wallace O.; Weir A. An analysis of the attrition of drug candidates from four major pharmaceutical companies. Nat. Rev. Drug Discovery 2015, 14, 475–486. 10.1038/nrd4609. - DOI - PubMed
    1. Russakovsky O.; Deng J.; Su H.; Krause J.; Satheesh S.; Ma S.; Huang Z.; Karpathy A.; Khosla A.; Bernstein M.; Berg A. C.; Fei-Fei L. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comp. Vis (IJCV) 2015, 115, 211–252. 10.1007/s11263-015-0816-y. - DOI
    1. Deng L.; Hinton G.; Kingsbury B. New types of deep neural network learning for speech recognition and related applications: An overview. Int. Conf. Acous. Speech Signal Proc. 2013, 8599–8603. 10.1109/ICASSP.2013.6639344. - DOI