Abstract
Artificial intelligence (AI) receives attention in media as well as in academe and business. In media coverage and reporting, AI is predominantly described in contrasted terms, either as the ultimate solution to all human problems or the ultimate threat to all human existence. In academe, the focus of computer scientists is on developing systems that function, whereas philosophy scholars theorize about the implications of this functionality for human life. In the interface between technology and philosophy there is, however, one imperative aspect of AI yet to be articulated: how do intelligent systems make inferences? We use the overarching concept “Artificial Intelligent Behaviour” which would include both cognition/processing and judgment/behaviour. We argue that due to the complexity and opacity of artificial inference, one needs to initiate systematic empirical studies of artificial intelligent behavior similar to what has previously been done to study human cognition, judgment and decision making. This will provide valid knowledge, outside of what current computer science methods can offer, about the judgments and decisions made by intelligent systems. Moreover, outside academe—in the public as well as the private sector—expertise in epistemology, critical thinking and reasoning are crucial to ensure human oversight of the artificial intelligent judgments and decisions that are made, because only competent human insight into AI-inference processes will ensure accountability. Such insights require systematic studies of AI-behaviour founded on the natural sciences and philosophy, as well as the employment of methodologies from the cognitive and behavioral sciences.
Similar content being viewed by others
Notes
Knight, Will (14 March 2017). “DARPA is funding projects that will try to open up AI's black boxes”. MIT Technology Review. https://www.technologyreview.com/s/603795/the-us-military-wants-its-autonomous-machines-to-explain-themselves/.
Sample, Ian (5 November 2017). “Computer says no: why making AIs fair, accountable and transparent is crucial”. the Guardian. https://www.theguardian.com/science/2017/nov/05/computer-says-no-why-making-ais-fair-accountable-and-transparent-is-crucial.
The EPSRC ‘Human-Like Computing’ initiative aims to bridge this ‘gap’ between ‘symbolic’/’rational’ and ‘neural’/’empirical’ AI. See: http://hlc.doc.ic.ac.uk/.
Cfr stimulus—response and classical conditioning.
New York City Council (2018). A Local Law in relation to automated decision systems used by agencies. http://legistar.council.nyc.gov/LegislationDetail.aspx?ID=3137815&GUID=437A6A6D-62E1-47E2-9C42-461253F9C6D0.
EU Parliament (2016). EU Framework on algorithmic accountability and transparency. http://www.europarl.europa.eu/sides/getDoc.do?pubRef=-//EP//TEXT+WQ+E-2016-007674+0+DOC+XML+V0//EN.
References
Ancona D, Bono V, Bravetti M, Campos J, Castagna G, Deniélou PM et al (2016) Behavioral types in programming languages. Found Trends Program Lang 3(2–3):95–230
Anderson FH (1948) The philosophy of Francis Bacon. University of Chicago Press, Chicago
Andrews GR, Schneider FB (1983) Concepts and notations for concurrent programming. ACM Comput Surv (CSUR) 15(1):3–43
Bäck T, Fogel DB, Michalewicz Z (1997) Handbook of evolutionary computation. CRC Press, Boca Raton
Bacon F, Montagu B (1857) The works of Francis Bacon, vol 1. Parry & McMillan, Philadelphia
Baeza-Yates R (2018) Bias on the web. Commun ACM 61(6):54–61
Ben-Ari M (2006) Principles of concurrent and distributed programming. Pearson Education, London
Bergstra JA, Ponse A, Smolka SA (eds) (2001) Handbook of process algebra. Elsevier, Amsterdam
Biamonte J, Wittek P, Pancotti N, Rebentrost P, Wiebe N, Lloyd S (2017) Quantum machine learning. Nature 549(7671):195
Binns R (2018) Algorithmic accountability and public reason. Philos Technol 31:543–556
Biran O, Cotton C (2017) Explanation and justification in machine learning: a survey. In: IJCAI-17 workshop on explainable AI (XAI), vol 8
Boden MA (2016) AI: its nature and future. Oxford University Press, Oxford
Bostrom N (2014) Superintelligence: paths, dangers, strategies. Oxford University Press, Oxford
Bottou L (2014) From machine learning to machine reasoning. Mach Learn 94(2):133–149
Brundage M, Avin S, Clark J, Toner H, Eckersley P, Garfinkel B, Dafoe A, Scharre P, Zeitzoff T, Filar B, Anderson H (2018) The malicious use of artificial intelligence: forecasting, prevention, and mitigation. arXiv preprint arXiv:1802.07228
Burling R (2005) The talking ape: how language evolved, vol 5. Oxford University Press, New York
Coello CAC, Lamont GB, Van Veldhuizen DA (2007) Evolutionary algorithms for solving multi-objective problems, 2nd edn. Springer, Berlin
Coffey P (1914) Ontology, or, the theory of being: an introduction to general metaphysics, vol 25. Longmans, London
Coffey P (1917) Epistemology; or, the theory of knowledge: an introduction to general metaphysics, vol 1. Longmans, London
Crabtree A, Urquhart L, Chen J (2019) Right to an explanation considered harmful. Edinburgh School of Law Research Paper Forthcoming. Available at SSRN: https://ssrn.com/abstract=3384790 or http://dx.doi.org/10.2139/ssrn.3384790
Dahl ES (2018) Appraising black-boxed technology: the positive prospects. Philos Technol 31:571–591
Danaher J (2016) The threat of algocracy: reality, resistance and accommodation. Philos Technol 29(3):245–268
De Laat PB (2018) Algorithmic decision-making based on machine learning from big data: can transparency restore accountability? Philos Technol 31:525–541
Demuth HB, Beale MH, De Jess O, Hagan MT (2014) Neural network design. Martin Hagan, Boston
Dennett D (1995) Darwin’s dangerous idea: evolution and the meanings of life. Simon & Schuster, New York
Dijkstra E (1965) Solution of a problem in concurrent programming control. Commun ACM 8(9):569
Dix A (1992) Human issues in the use of pattern recognition techniques. In: Beale R, Finlay J (eds) Neural networks and pattern recognition in human computer interaction. Ellis Horwood, Chichester, pp 429–451
Dix A (2016) Human-like computing. (Personal report on the EPRC workshop of the same name) http://alandix.com/blog/2016/02/23/human-like-computing/
Dix A (2018) Sufficient Reason. Keynote at HCD for Intelligent Environments, BHCI, Belfast, 3rd July 2018. http://alandix.com/academic/talks/sufficient-reason-2018/
Doshi-Velez F, Kim B (2017) Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608
Dwork C (2011) A firm foundation for private data analysis. Commun ACM 54(1):86–95
Dylan H, Goodman MS, Jackson P, Jansen PT, Maiolo J, Pedersen T (2017) The way of the Norse Ravens: merging profession and academe in Norwegian national intelligence higher education. Intell National Secur 32(7):944–960
Ferrucci D (2011) How it all began and what’s next. IBM Research. https://www.ibm.com/blogs/research/2011/12/dave-ferrucci-at-computer-history-museum-how-it-all-began-and-whats-next/
Ferrucci D, Levas A, Bagchi S, Gondek D, Mueller ET (2013) Watson: beyond jeopardy! Artif Intell 199:93–105
Frégnac Y (2017) Big data and the industrialization of neuroscience: a safe roadmap for understanding the brain? Science 358(6362):470–477
Frensch PA, Runger D (2003) Implicit learning. Curr Dir Psychol Sci 12:13–18
Gärling T, Ettema D, Friman M (eds) (2014) Handbook of sustainable travel. Springer, Berlin
Gibson JJ (1979) The Ecological Approach to visual perception. Houghton-Mifflin, Boston
Gilovich T, Griffin D, Kahneman D (2002) Heuristics and biases: the psychology of intuitive judgment. Cambridge University Press, Cambridge
Gordon AD, Henzinger TA, Nori AV, Rajamani SK (2014) Probabilistic programming. In: Proceedings of the 36th international conference on software engineering (ICSE)—future of software engineering track. ACM, pp 167–181
Hennessy M (1988) Algebraic theory of processes. MIT press, Cambridge
Hennessy M, Rathke J (2004) Typed behavioural equivalences for processes in the presence of subtyping. Math Struct Comput Sci 14(5):651–684
Holland JH (1992) Adaptation in natural and artificial systems: an introductory analysis with applications to biology, control, and artificial intelligence. MIT press, Cambridge
Hurlburt RT, Knapp TJ (2006) Münsterberg in 1898, not Allport in 1937, introduced the terms ‘idiographic’ and ‘nomothetic’ to American psychology. Theory Psychol 16(2):287–293
Jackson S, Jacobs S (1980) Structure of conversational argument: pragmatic bases for the enthymeme. Q J Speech 66(3):251–265
Jordan MI, Mitchell TM (2015) Machine learning: trends, perspectives, and prospects. Science 349(6245):255–260
Kahneman D (2003) A perspective on judgment and choice: mapping bounded rationality. Am Psychol 58:697–720
Kahneman D, Thaler RH (2006) Anomalies: utility maximization and experienced utility. J Econ Perspect 20(1):221–234
Kahneman D, Tversky A (1972) Subjective probability: a judgment of representativeness. Cogn Psychol 3(3):430–454
Katoen JP, Gretz F, Jansen N, Kaminski BL, Olmedo F (2015) Understanding probabilistic programs. Correct system design—symposium in honor of Ernst-Rüdiger Olderog (vol 9360 of lecture notes in computer science. Springer, Cham, pp 15–32
Kirkham RL, Kirkham RL (1992) Theories of truth: a critical introduction (No. s 401). MIT press, Cambridge
Kneale W (1945). Truths of logic. In: Proceedings of the aristotelian society, vol 46. Aristotelian Society, Wiley, London, pp 207–234
Knuth DE (1973) The art of computer programming vol. 1: fundamental algorithms, 2nd edn. Addison-Wesley Publishing, Boston
Kuhn TS (1962) The structure of scientific revolutions. University of Chicago Press, Chicago
Lazer D, Kennedy R, King G, Vespignani A (2014) The parable of Google Flu: traps in big data analysis. Science 343(6176):1203–1205
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436
McCarthy J (1987) Generality in artificial intelligence. Commun ACM 30(12):1030–1035
McCarthy J, Hayes PJ (1969) Some philosophical problems from the standpoint of artificial intelligence. In: Webber BL, Nilsson NJ (eds) Readings in artificial intelligence (1981). Morgan Kaufmann Publishers, Los Altos, California, pp 431–450
McLuhan M (1964) Understanding media: the extensions of man. MIT Press, USA
Miller T, Howe P, Sonenberg L (2017) Explainable AI: beware of inmates running the asylum or: how I learnt to stop worrying and love the social and behavioural sciences. arXiv preprint arXiv:1712.00547
Minsky ML (1967) Computation: finite and infinite machines. Prentice-Hall, Englewood Cliffs
Mitchell M (1998) An introduction to genetic algorithms. MIT press, Cambridge
Monroe D (2018) AI explain yourself. Commun ACM 61(11):11–13. https://doi.org/10.1145/3276742
Murdoch WJ, Singh C, Kumbier K, Abbasi-Asl R, Yu B (2019) Interpretable machine learning: definitions, methods, and applications. arXiv preprint arXiv:1901.04592
Newell A, Simon HA (1972) Human problem solving, vol 104, No. 9. Prentice-Hall, Englewood Cliffs
Newell A, Simon HA (1976) Computer science as empirical inquiry: symbols and search. Commun ACM 19(3):113–126
Newell A, Shaw JC, Simon HA (1958) Elements of a theory of human problem solving. Psychol Rev 65(3):151. https://doi.org/10.1037/h0048495
Osawa E, Kitano H, Asada M, Kuniyoshi Y, Noda I (1996) RoboCup: the robot world cup initiative. In: Proceedings of the second international conference on multi-agent systems (ICMAS), Kyoto, Japan, pp 9–13
Pagallo U (2018) Algo-rhythms and the beat of the legal drum. Philos Technol 31:507–524
Parkes DC, Wellman MP (2015) Economic reasoning and artificial intelligence. Science 349(6245):267–272
Pedersen T, Friman M, Kristensson P (2011) Affective forecasting: predicting and experiencing satisfaction with public transportation. J Appl Soc Psychol 41(8):1926–1946
Pedersen T, Johansen C, Jøsang A (2018) Behavioural computer science: an agenda for combining modelling of human and system behaviours. Hum-centric Comput Inf Sci 8(1):7
Popper KR (1972) Objective knowledge: an evolutionary approach. Oxford University Press, Oxford
Postman N (1992) Technopoly: the surrender of culture to technology. Knopf, New York, p 1992
Rahwan I, Simari GR (2009) Argumentation in artificial intelligence. Springer, Berlin
Ramprasad R, Batra R, Pilania G, Mannodi-Kanakkithodi A, Kim C (2017) Machine learning in materials informatics: recent applications and prospects. npj Comput Mater 3(1):54
Rescher N (1973) The coherence theory of truth. Clarendon Press, Oxford, pp 54–64
Robinson OC (2011) The idiographic/nomothetic dichotomy: tracing historical origins of contemporary confusions. Hist Philos Psychol 13(2):32–39
Russell SJ (1997) Rationality and intelligence. Artif Intell 94(1–2):57–77
Russell SJ, Norvig P (2016) Artificial intelligence: a modern approach. Pearson Education Limited, London
Samek W, Wiegand T, Müller KR (2017) Explainable artificial intelligence: understanding, visualizing and interpreting deep learning models. arXiv preprint arXiv:1708.08296
Simon HA (1979) Rational decision making in business organizations (Nobel Prize lecture). Am Econ Rev 69(4):493–513
Simon HA (1996) The sciences of the artificial. MIT press, Cambridge
Sloman SA (2002) Two systems of reasoning. In: Gilovich T, Griffin D, Kahneman D (eds) Heuristics and biases: the psychology of intuitive judgment. Cambridge University Press, Cambridge
Smiley TJ (1973) What is a syllogism? J Philos Log 2(1):136–154
Sutton RS, Barto AG (2018) Reinforcement learning: an introduction, 2nd edn. MIT press, Cambridge
Tversky A, Kahneman D (1973) Availability: a heuristic for judging frequency and probability. Cogn Psychol 5(2):207–232
Tversky A, Kahneman D (1974) Judgment under uncertainty: heuristics and Biases. Science 185(4157):1124–1131
USACM: US Public Policy Council of The Association for Computing Machinery (2017) Statement on algorithmic transparency and accountability. Association for Computing Machinery, New York
Vardi M (2018) Machine learning and logic: fast and slow thinking. Invited talk at the summit on machine learning meets formal methods. https://easychair.org/smart-program/FLoC2018/SoMLMFM-2018-07-13.html#talk:76996
Wachter S, Mittelstadt B, Floridi L (2017) Transparent, explainable, and accountable AI for robotics. Sci Robot 2(6):eaan6080
Walton D (2014) Abductive reasoning. University of Alabama Press, Tuscaloosa
Zemel R, Wu Y, Swersky K, Pitassi T, Dwork C (2013) Learning fair representations. In: International conference on machine learning, pp 325–333
Acknowledgements
We are very grateful to the anonymous reviewers for their valuable comments that have improved this paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Tore Pedersen was partially supported by the project Oslo Analytics funded by the IKTPLUSS program of the Norwegian Research Council. Christian Johansen was partially supported by the project IoTSec—Security in IoT for Smart Grids, with Number 248113/O70 part of the IKTPLUSS program funded by the Norwegian Research Council.
Rights and permissions
About this article
Cite this article
Pedersen, T., Johansen, C. Behavioural artificial intelligence: an agenda for systematic empirical studies of artificial inference. AI & Soc 35, 519–532 (2020). https://doi.org/10.1007/s00146-019-00928-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00146-019-00928-5