Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2020 Nov 20;3(1):27.
doi: 10.1186/s42492-020-00063-9.

Modeling of moral decisions with deep learning

Affiliations

Modeling of moral decisions with deep learning

Christopher Wiedeman et al. Vis Comput Ind Biomed Art. .

Abstract

One example of an artificial intelligence ethical dilemma is the autonomous vehicle situation presented by Massachusetts Institute of Technology researchers in the Moral Machine Experiment. To solve such dilemmas, the MIT researchers used a classic statistical method known as the hierarchical Bayesian (HB) model. This paper builds upon previous work for modeling moral decision making, applies a deep learning method to learn human ethics in this context, and compares it to the HB approach. These methods were tested to predict moral decisions of simulated populations of Moral Machine participants. Overall, test results indicate that deep neural networks can be effective in learning the group morality of a population through observation, and outperform the Bayesian model in the cases of model mismatches.

Keywords: Artificial intelligence; Bayesian method; Deep learning; Moral machine experiment.

PubMed Disclaimer

Conflict of interest statement

The authors declare that they have no competing interests.

Figures

Fig. 1
Fig. 1
Moral Machine Example: Example screenshot of a scenario from a Moral Machine scenario [8]. The participant must decide the more ethical course of action for the self-driving vehicle: swerving would kill all vehicle passengers (right), while maintaining the course would result in the death of all pedestrians (left) [8]
Fig. 2
Fig. 2
Feature Transform: Binary transformation matrix A used in Kim, et al., which converts a set of character traits θ into quantifiable features [9]
Fig. 3
Fig. 3
Moral Principle Mean: Mean value for all underlying distributions of the moral principle vector w
Fig. 4
Fig. 4
Moral Principle Covariance: Covariance matrix for all underlying distributions of the moral principle vector w
Fig. 5
Fig. 5
Distribution Transform: Histograms of marginal distributions of the ’Human’ value for the normal dataset (left) and one of the transformed datasets (right)
Fig. 6
Fig. 6
Neural Network Architecture: Network architecture for the DL moral decision model. Batchnorm signifies a batch normalization layer
Fig. 7
Fig. 7
Generated Distribution Mean: Mean values of the principle vector w used to synthesize more general data
Fig. 8
Fig. 8
Generated Distribution Covariance: Covariance matrix of the principle vector w used to synthesize more general data
Fig. 9
Fig. 9
DL Accuracy: DL model predictive accuracy with different sample sizes and underlying distributions of w. Various distributions for w are denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution)
Fig. 10
Fig. 10
DL Accuracy (Deterministic Decision Process): DL model predictive accuracy with different sample sizes and underlying distributions of w. Various distributions for w are denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution) In this test, decisions were simulated with the deterministic version of Eq. 2
Fig. 11
Fig. 11
Model Performances: Comparison of model performances (DL trained with 2,000 participants involved) over datasets with various underlying distributions of w, denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution). ‘GT’ denotes the predictive model in which the exact w for each participant is known
Fig. 12
Fig. 12
Model Performances (Deterministic Decision Process): Comparison of the model performances (DL trained with 2,000 participants involved) over datasets with various underlying distributions of w, denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution). In this test, decisions were simulated with the deterministic version of Eq. 2
Fig. 13
Fig. 13
DL Model Accuracy (Generated): DL model predictive accuracy with different sample sizes and underlying distributions of w, applied to the generalized, abstract data. Various distributions for w are denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution)
Fig. 14
Fig. 14
DL Model Accuracy (Generated; Deterministic Process): DL model predictive accuracy with different sample sizes and underlying distributions of w, applied to the generalized, abstract data. Various distributions for w are denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution) In this test, decisions were simulated with deterministic version of Eq. 2
Fig. 15
Fig. 15
Model Performances (Generated): Comparison of model performances (DL trained with 2,000 participants involved) over datasets with various underlying distributions of w, denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution), as applied to the generalized, abstract data. ‘GT’ is the predictive model in which the exact w for each test participant is known
Fig. 16
Fig. 16
Model Performances (Generated; Deterministic Process): Comparison of model performances (DL trained with 2,000 participants involved) over datasets with various underlying distributions of w, denoted by the approximate average of absolute shape factors k (k¯=0 denotes the Gaussian distribution), as applied to the generalized, abstract data. In this test, decisions were simulated with the deterministic version of Eq. 2

References

    1. Bigman YE, Gray K. People are averse to machines making moral decisions. Cognition. 2018;181:21–34. doi: 10.1016/j.cognition.2018.08.003. - DOI - PubMed
    1. Kramer MF, Schaich Borg J, Conitzer V, Sinnott-Armstrong W. Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society. AIES ’18. New York, NY, USA: Association for Computing Machinery; 2018. When do people want ai to make decisions? pp. 204–209.
    1. Marcus G (2020) The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence. arXiv:2002.06177. https://arxiv.org/abs/2002.06177.
    1. Shaw NP, Stöckel A, Orr RW, Lidbetter TF, Cohen R (2018) Towards provably moral ai agents in bottom-up learning frameworks. AIES ’18, 271–277.. Association for Computing Machinery, New York, NY, USA.
    1. Bentham J. An introduction to the principles of morals and legislation. Oxford: Oxford University Press; 1789.

LinkOut - more resources