Artificial Intelligence Course Book PDF

Summary

This course book provides an introduction to Artificial Intelligence (AI), covering its history, early systems, modern systems, and various applications. It details essential readings and learning objectives for a course on AI. The book aims to provide a foundational understanding of AI concepts for students.

Full Transcript

ARTIFICIAL INTELLIGENCE DLMAIAI01 ARTIFICIAL INTELLIGENCE MASTHEAD Publisher: IU Internationale Hochschule GmbH IU International University of Applied Sciences Juri-Gagarin-Ring 152 D-99084 Erfurt Mailing address: Albert-Proeller-Straße 15-19 D-86675 Buchdorf...

ARTIFICIAL INTELLIGENCE DLMAIAI01 ARTIFICIAL INTELLIGENCE MASTHEAD Publisher: IU Internationale Hochschule GmbH IU International University of Applied Sciences Juri-Gagarin-Ring 152 D-99084 Erfurt Mailing address: Albert-Proeller-Straße 15-19 D-86675 Buchdorf [email protected] www.iu.de DLMAIAI01 Version No.: 002-2024-0510 N. N. Created with Midjourney on behalf of IU, 2024, using the prompt: "the human brain and artificial intelligence in modern computers, cpu icon, abstract, futuristic, cyberspace --ar 16:9 --v 6.0" © 2024 IU Internationale Hochschule GmbH This course book is protected by copyright. All rights reserved. This course book may not be reproduced and/or electronically edited, duplicated, or dis- tributed in any kind of form without written permission by the IU Internationale Hoch- schul GmbH (hereinafter referred to as IU). The authors/publishers have identified the authors and sources of all graphics to the best of their abilities. However, if any erroneous information has been provided, please notify us accordingly. 2 TABLE OF CONTENTS ARTIFICIAL INTELLIGENCE Introduction Signposts Throughout the Course Book............................................. 6 Basic Reading.................................................................... 7 Required Reading................................................................. 8 Further Reading................................................................. 10 Learning Objectives.............................................................. 12 Unit 1 History of Artificial Intelligence 13 1.1 Historical Developments...................................................... 14 1.2 AI Winters................................................................... 23 1.3 Notable Advances in Artificial Intelligence...................................... 26 Unit 2 Early Systems in Artificial Intelligence 31 2.1 Overview of Expert Systems................................................... 32 2.2 Introduction to Prolog........................................................ 34 2.3 Pattern Recognition and Machine Learning (ML)................................ 36 2.4 Use Cases................................................................... 37 Unit 3 Neuroscience and Cognitive Science 41 3.1 Neuroscience and the Human Brain........................................... 42 3.2 Cognitive Science............................................................ 46 3.3 The Relationship Between Neuroscience, Cognitive Science, and Artificial Intelligence................................................................................. 48 Unit 4 Modern Artificial Intelligence Systems 57 4.1 Recent Developments in Hardware and Software............................... 58 4.2 Narrow Versus General Artificial Intelligence.................................... 64 4.3 Natural Language Processing (NLP) and Computer Vision........................ 65 3 Unit 5 Applications of Artificial Intelligence 73 5.1 Mobility and Autonomous Vehicles............................................ 74 5.2 Personalized Medicine....................................................... 77 5.3 FinTech..................................................................... 81 5.4 Retail and Industry........................................................... 86 Appendix List of References................................................................ 96 List of Tables and Figures........................................................ 100 4 INTRODUCTION WELCOME SIGNPOSTS THROUGHOUT THE COURSE BOOK This course book contains the core content for this course. Additional learning materials can be found on the learning platform, but this course book should form the basis for your learning. The content of this course book is divided into units, which are divided further into sec- tions. Each section contains only one new key concept to allow you to quickly and effi- ciently add new learning material to your existing knowledge. At the end of each section of the digital course book, you will find self-check questions. These questions are designed to help you check whether you have understood the con- cepts in each section. For all modules with a final exam, you must complete the knowledge tests on the learning platform. You will pass the knowledge test for each unit when you answer at least 80% of the questions correctly. When you have passed the knowledge tests for all the units, the course is considered fin- ished and you will be able to register for the final assessment. Please ensure that you com- plete the evaluation prior to registering for the assessment. Good luck! 6 BASIC READING Russell, S & Norvig, P. (2022). Artificial intelligence. A modern approach (4th ed.). Pearson Education. https://search.ebscohost.com/login.aspx?direct=true&AuthType=sso&db= cat09158a&AN=iuo.oai.edge.iu.folio.ebsco.com.fs00001148.aefe7710.a82e.54fa.82e3.c 720acab8c2b&lang=de&site=eds-live&scope=site&custid=s6068579 Chowdhary, K. R. (2020). Fundamentals of Artificial Intelligence. Springer India. http://searc h.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=true&db=edsbvb&AN=edsbvb.B V046660982&site=eds-live&scope=site Ward, J. (2020). The student's guide to cognitive neuroscience. (4th ed.). Taylor & Francis Group. http://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=true&db=cat 05114a&AN=ihb.45699&site=eds-live&scope=site 7 REQUIRED READING UNIT 1 Russell, S. & Norvig, P. (2022). Artificial intelligence: A modern approach (4th ed.). Pearson Education. Chapter 1. https://search.ebscohost.com/login.aspx?direct=true&AuthTyp e=sso&db=cat09158a&AN=iuo.oai.edge.iu.folio.ebsco.com.fs00001148.aefe7710.a82e. 54fa.82e3.c720acab8c2b&lang=de&site=eds-live&scope=site&custid=s6068579 Bringsjord, S., & Govindarajulu, N.S. (2022). Artificial intelligence. In Zalta, E.N. (Ed.).The Stanford Encyclopedia of Philosophy (Fall 2022 Edition). https://plato.stanford.edu/arc hives/fall2022/entries/artificial-intelligence/ UNIT 2 Lucas, P.J.F. & Van der Gaag, L. (1991). Principles of expert systems. Amsterdam: Addison Wesley (copyright returned to author). Chapters 1-3 (Available online) Chowdhary, K. R. (2020). Fundamentals of artificial intelligence. Springer India. Chapter 5. h ttp://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=true&db=edsbvb&AN =edsbvb.BV046660982&site=eds-live&scope=site UNIT 3 Ward, J. (2020). The student's guide to cognitive neuroscience. (4th ed.). Taylor & Francis Group. Chapter 1. http://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=tr ue&db=cat05114a&AN=ihb.45699&site=eds-live&scope=site Frankish, K. & Ramsey, W.M. (Eds.) (2012). The Cambridge handbook of cognitive science. Cambridge: Cambridge University Press. Part I – Foundations. http://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=true&db=edsbas&AN=edsbas.12E504D2&sit e=eds-live&scope=site UNIT 4 Mell, P., & Grance, T. (2011). The NIST definition of cloud computing: Recommendations of the National Institute of Standards and Technology. National Institute of Standards and Technology. (Available online). Russell, S. & Norvig, P. (2022). Artificial intelligence: A modern approach (4th ed.). Pearson Education. Chapters 24, 27. https://search.ebscohost.com/login.aspx?direct=true&Aut hType=sso&db=cat09158a&AN=iuo.oai.edge.iu.folio.ebsco.com.fs00001148.aefe7710. a82e.54fa.82e3.c720acab8c2b&lang=de&site=eds-live&scope=site&custid=s6068579 8 UNIT 5 Marr, B. (2019). Artificial intelligence in practice: How 50 successful companies used AI and machine learning to solve problems. John Wiley & Sons. Chapters 18, 20, 31, 32, 37, 49. h ttp://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=true&db=cat05114a& AN=ihb.45473&site=eds-live&scope=site Lyman, K. T. (2020). A practical introduction to artificial intelligence in medicine. In L. Gold- schmidt & R. M. Relova (Eds.), Patient-centered digital healthcare technology: Novel applications for next generation healthcare systems (pp. 243–268). Institution of Engi- neering and Technology (The IET). http://search.ebscohost.com.pxz.iubh.de:8080/logi n.aspx?direct=true&db=edsair&AN=edsair.doi...........03a20b6f722c6b43312ffc5d088e0 f54&site=eds-live&scope=site 9 FURTHER READING UNIT 1 Ford, M. (2018). Architects of intelligence: The truth about AI from the people building it. Packt Publishing. http://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=tr ue&db=cat05114a&AN=ihb.45667&site=eds-live&scope=site UNIT 2 Clocksin, W.F. & Mellish, C.S. (2003). Programming in Prolog (5th ed.). Springer. http://searc h.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct=true&db=edsair&AN=edsair.doi...........bdea32b1f153a656183f68f8d544da69&site=eds-live&scope=site Körner, P., Leuschel, M., Barbosa, J., Costa, V. S., Dahl, V., Hermenegildo, M. V., Morales, J. F., Wielemaker, J., Diaz, D., Abreu, S., & Ciatto, G. (2022). Fifty years of Prolog and beyond: Theory and practice of logic programming, 1–83. (Available online) UNIT 3 Frankish, K. & Ramsey, W.M. (Eds.) (2012). The Cambridge handbook of cognitive science. Cambridge University Press. Part II – Aspects of Cognition http://search.ebscohost.co m.pxz.iubh.de:8080/login.aspx?direct=true&db=edsbas&AN=edsbas.12E504D2&site=e ds-live&scope=site UNIT 4 Jurafsky, D., Martin, J.H. (2022). Speech and language processing: An introduction to natu- ral language processing, computational linguistics, and speech recognition (3rd ed.) (Available Online) Davies, E.R. (2018). Computer vision: Principles, algorithms, applications, learning. (5th ed.). Academic Press. Part 1: Low level vision http://search.ebscohost.com.pxz.iubh.de:808 0/login.aspx?direct=true&db=edsebk&AN=1204289&site=eds-live&scope=site UNIT 5 Davenport, T.H. (2018). The AI advantage: How to put the artificial intelligence revolution to work. The MIT Press. http://search.ebscohost.com.pxz.iubh.de:8080/login.aspx?direct =true&db=edsoai&AN=edsoai.on1268862475&site=eds-live&scope=site 10 McCreadie, R., Perakis, K., Srikrishna, M., Droukas, N., Pitsios, S., Prokopaki, G., Perdi- kouri, E., Macdonald, C., & Ounis, I. (2022). Next-generation personalized investment recommendations. In J. Soldatos & D. Kyriazis (Eds.), Big data and artificial intelligence in digital finance: Increasing personalization and trust in digital finance using big data and AI (pp. 171–198). Springer International Publishing. http://search.ebscohost.com. pxz.iubh.de:8080/login.aspx?direct=true&db=edssjb&AN=edssjb.978.3.030.94590.9.10 &site=eds-live&scope=site 11 LEARNING OBJECTIVES Welcome to Artificial Intelligence. This course will provide you with an introduction to artificial intelligence. To this end, we will revisit important events in the history of the field to understand its origins as well as important trends and paradigms that have shaped it. As a discipline, artificial intelligence draws upon a multitude of ideas that originate from related fields of study, such as neuroscience and cognitive science. Consequently, an overview of these areas will be given alongside an exposition of how they relate to artificial intelligence research and methodology. Furthermore, you will be introduced to contemporary cur- rents and issues in the study of artificial intelligence and its applications. A significant part of your personal future is your career, which will very likely be impacted by artificial intelligence, requiring new skills for which demand currently outstrips supply. Congratulations on taking the first step! 12 UNIT 1 HISTORY OF ARTIFICIAL INTELLIGENCE STUDY GOALS On completion of this unit, you will have learned … – how artificial intelligence has developed as a scientific discipline. – what paradigms have dominated public perception of the field at different times. – which notable advances are still relevant today. – what the history of artificial intelligence means for you in terms of learning new skills and contributing to society. 1. HISTORY OF ARTIFICIAL INTELLIGENCE Introduction This unit starts by providing an overview of artificial intelligence and then proceeds to explain some of the details that make up this new technology. It will take you back to the year 350 BC to illustrate how ideas about the artificial creation of intelligent processes have evolved across the ages. To date, artificial intelligence is an active research area in the field of cognitive psychology. Consequently, no single uncontroversial or uncontested definition of the term currently exists. An operational definition that is apposite with respect to the subject of this course can be stated as follows: intelligence is the ability to accomplish complex goals. In this vein, artificial intelligence (AI) can be framed as the totality of abilities used to accomplish complex goals that are achieved by machines. This broad definition implies perception, learning, problem solving, and decision-making. In practical terms, this means that com- puters that can perceive via sensing devices, apply reasoning, and then act in a rational manner are defined as exhibiting artificial intelligence. 1.1 Historical Developments Ancient Artificial Intelligence History A conventional starting point for the history of artificial intelligence tends to be around the 1950s when the notion of artificial intelligence was first applied to the nascent discipline of computer science. However, it is enlightening to take a broader view of the history of ideas related to artificial intelligence through a brief look at some of the key thinkers in the field dating back to 350 BC and the philosophical insights of Aristotle (Russell & Norvig, 2022). Aristotle (384—322 BC), Greek Philosophical insight: Artistotle formalized logical conclusions by fully enumerating all possible categorical syllogisms. Relation to artificial intelligence: Algorithms can be programmed to derive valid logical conclusions based on a given set of rules. Syllogism Aristotle taught logic at a school called the Lyceum. Syllogisms are rules for arriving at This is a form of logical workable conclusions from two or more propositions. A contemporary equivalent to the argumentation that applies deductive reason- enumeration of valid forms of logical derivation forms the foundation of logical program- ing to reach a conclusion ming languages. based on two or more propositions that are assumed to be true. 14 Leonardo da Vinci (1452—1519), Italian Philosophical insight: Da Vinci designed a hypothetical computing machine on paper. Relation to artificial intelligence: Progress in computing machinery is a necessary pre-con- dition for artificial intelligence. Da Vinci designed a computing machine with 13 registers, demonstrating that a black box can accept inputs and produce outputs based on a stored program in memory or mechan- ics. Thomas Hobbes (1588—1679), British Philosophical Insight: Hobbes derived the notion of the necessity of a social contract from a mechanistic understanding of the individual. Relation to artificial intelligence: Hobbes identified similarities between reasoning and computation in that humans employ operations similar to calculus in rational decision- making, such that they could be formalized, analogous to mathematics. René Descartes (1596—1650), French Philosophical Insight: Rationality and reason can be defined via mechanics and mathe- matics. Similar views on rational thought were held by Gottfried Wilhelm Leibniz (1646— 1716) based on earlier work by Ramon Llull (1232—1315). Relation to artificial intelligence: Objectives can be formulated in the form of equations. Objectives in linear programming or in artificial intelligence agents are defined mathemat- ically. Descartes described rationalism and materialism as two sides of the same coin. Arti- ficial intelligence aims for rational decisions arrived at mathematically. David Hume (1711—1776), Scottish Philosophical Insight: Hume completed foundational work on questions of logical induc- tion and the notion of causation. Relation to artificial intelligence: Hume linked the principles of learning with repeated exposure that, among others, manifests in the learning curve. The principle of inferring Learning curve patterns or relations in data via repeated exposure is key to many machine learning algo- This is a graphical repre- sentation showing how rithms. an increase in learning (measured on the vertical Recent Artificial Intelligence History axis) derives from greater experience (the horizon- tal axis). Recent history in artificial intelligence is commonly said to have started in 1956, the year of the seminal Dartmouth conference. The term artificial intelligence was first coined at this conference and a tentative definition of the concept proposed. We deal with this recent history in terms of key personalities, key organizations that supported the cause, and key concepts that broke previous limits. 15 Dartmouth conference Key personalities The Summer Research Project on Artificial Intelli- gence was held in 1956 at A recent history of artificial intelligence often begins by describing the contributions of Dartmouth College in individuals such as Alan Turing (1912—1954), John McCarthy (1927—2011), Marvin Minsky New Hampshire, USA. (1927—2016), and Noam Chomsky (1928—). The female programmers of the ENIAC (Elec- This six-week conference is considered to be the tronic Numerical Integrator and Computer), the first general-purpose computer, which birthplace of artificial began operation in 1946, should also be mentioned in this context (Schwartz, 2019). These intelligence as a research are notably, Kathleen McNulty, Frances Bilas, Betty Jean Jennings, Ruth Lichterman, Eliza- discipline. beth Snyder, and Marlyn Wescoff. Alan Turing was an English mathematician and computer scientist who became fascinated with the prospect of mechanizing and formalizing rational thought processes. It was Turing who famously came up with the eponymous Turing test in 1950. In this test, a machine is seen to exhibit intelligence if it is able to have a conversation with a human observer that the latter cannot distinguish from a conversation with another human. Automata John McCarthy was an American scientist who studied automata and first coined the term This refers to a self-oper- artificial intelligence. Together with IBM and the Massachusetts Institute of Technology ating machine or machine control mechanism (MIT), he made artificial intelligence a separate field of study. McCarthy is also credited designed to automatically with inventing the Lisp programming language in 1958, which was used in artificial intelli- follow a predetermined gence applications for 30 years in the areas of fraud detection, website design, and robot- sequence of operations or instructions. ics. In the 1960s, he also invented computer timesharing and founded the Stanford Artifi- cial Intelligence Laboratory, which played a pivotal role in research relating to human skills for machines, such as seeing, listening, and reasoning. Marvin Minsky was an early artificial intelligence researcher and cognitive scientist who, together with McCarthy, attended the first artificial intelligence conference at Dartmouth College. In 1959, they jointly founded the MIT Artificial Intelligence Laboratory. However, their collaboration ended after McCarthy moved to Stanford University, with Minsky remaining at MIT. Noam Chomsky is also worth a brief mention, more as a linguist and philosopher than as an artificial intelligence scientist. His contribution to artificial intelligence comes in the form of his criticism of social media and as a contributor to linguistics and cognition. Key institutions Key institutions involved in the development of artificial intelligence include universities such as Dartmouth College, the host of influential artificial intelligence conferences, and MIT, where many influential figures of early artificial intelligence research taught. Corpora- tions such as IBM and INTEL, and government research institutions such as the Defense Advanced Research Projects Agency (DARPA) in the United States, which funds basic dual technology research projects (i.e. projects with civilian as well as military purposes), have also played a critical role in the development of artificial intelligence. 16 Key ideas supporting the development of artificial intelligence Research in the areas of decision theory, game theory, neuroscience, and natural language processing have also contributed to the development of artificial intelligence. Decision theory combines probability (mathematics) and utility (economics) to frame artificial intelligence decisions in terms of economic benefit and uncertainty. Game theory was made famous by John von Neuman (1903—1957), an American-Hun- garian computer scientist, and Oskar Morgenstern (1902—1977), an American-German mathematician and game theorist. Their work led to rational agents learning strategies to solve games. Neuroscience is a body of knowledge about the brain that some artificial intelligence models try to emulate, especially the brain’s problem-solving and information storing abilities. Natural language processing (NLP) is a discipline at the confluence of linguistics and computer science that aims to analyze and process language in both written and spo- ken forms. High-level languages are closer to human language and allow programmers independ- ence from computing hardware’s instruction sets. Other artificial intelligence specific lan- guages are listed below: Lisp is one of the older computer programming languages developed by John McCar- thy. Its name derives from the words “list processing”. Lisp is uniquely able to process strings of characters. Although it dates back to the 1960s, it remains relevant today and was used for early artificial intelligence programming. Prolog is an early artificial intelligence programming language designed for solving logi- cal formulas and proving theorems. Python is a high-level, general purpose programming language that plays a large role in artificial intelligence today. It has been around since 1991. Being open-sourced, Python has a substantial library that coders (programmers) can use to create value quickly. Recent progress in artificial intelligence is linked to three main factors. The first is progress in the availability of massive amounts of data in many formats, called big data, which is Big data continuously increasing. The second is progress in the data processing capacity of com- This refers to datasets that exceed the capabili- puters, while the third relates to new insights gained through mathematics, philosophy, ties of conventional infor- cognitive science, and machine learning. mation processing tools, commonly characterized by the dimensions of vol- Key Trends in Artificial Intelligence ume, velocity, variety, and veracity. Some definitions also add value and valid- Investment in artificial intelligence is becoming exceedingly relevant for corporations, uni- ity. versities, and governments. There are hardly any aspects of society, the economy, and our own personal lives that are not already influenced by artificial intelligence technology. In particular, technology leaders, such as Apple, Amazon, Microsoft, Google, and important Chinese companies such as Baidu and Tencent are highly committed to research and development in artificial intelligence. Current developments can be classified into three major areas: 17 Assistive intelligence devices These include wearable health monitors that measure pulse, temperature, or blood sugar, and sensors used in automobiles that detect all kinds of mechanical and human-related conditions. These are then interpreted via artificial intelligence algorithms in order to pro- vide assistance services. Augmented intelligence In this field of research, artificial intelligence works alongside a human expert to provide task-related support in selected areas where machine performance is superior to human performance. The overall process is, however, driven by the human expert. An example is the assistance provided to physicians in prescribing medications to overcome the limita- tions of human medical knowledge by being able to consider all possible medical side effects in a few minutes. Another example is a warning signal given to a driver whose car is in the wrong lane. Autonomous intelligence This involves physical robotic actors, such as self-driving vehicles, moving warehouse robots, or the autonomous control of non-physical workflows and business processes. The overarching theme is that control is exerted, with decisions being made without human intervention. The prospect of artificial intelligence operating in the real world without human control is not without danger, posing important ethical questions. The reaction of autonomous vehicles when faced with life-and-death decisions is an obvious example. However, also seemingly more benign application domains such as the autonomous acceptance or denial of insurance claims is not without its perils considering the impact they may have on the economic prospects of the affected persons. The Future of Artificial Intelligence Assessing the future prospects or impact of a technology or field of research is always highly speculative and influenced by cognitive biases and previous experiences that may have lost their relevance for assessing the state of the field and its future prospects. Thus, no attempt is given here at predicting the long-term future of artificial intelligence. How- ever, it is prudent to examine a common pattern that emerges in the general uptake and development of innovations and how they relate to current topics in artificial intelligence and supporting technologies. Gartner (2018) captures the history of many innovations in the form of a hype curve. Hype curves are described in phases, as shown in the graph below. The x-axis represents time characterized by the following phases: a discovery phase, or a need in the marketplace, which triggers innovation a peak phase of inflated expectation, which is shorter than the other phases a period of disillusionment a period of enlightenment, which recognizes the value of an innovation a period of leveling off in which productivity takes hold and becomes the norm 18 Figure 1: Gartner Hype Cycle for Emerging Technologies, 2018 Source: Gartner, Svetlana Sicular and Kenneth Brant, 24 July 2018. 19 The technology hype curve bears some similarity to the bell-shaped curve of a normal dis- tribution with the exception that the right tail leads into a rising slope that eventually lev- els off. Several aspects of artificial intelligence are positioned along this curve. For exam- ple, on closer inspection we can observe the following trends: In the innovation trigger phase, knowledge graphs, neuromorphic hardware, AI PaaS, Artificial General Intelligence (the ability of a machine to perform humanlike intellectual tasks), Edge AI, and smart robots appear. At the peak phase of inflated expectations, or hype, we find deep neural networks that were the driver of new performance highs in many machine learning applications dur- ing the last five to seven years. In the disillusionment phase, that is going downhill, we find that Level 4 Autonomous Driving may experience defunding due to its failing to meet expectations. Nothing in the field of artificial intelligence has reached the plateau of productivity yet, which represents general acceptance and productive use of the technology. Governance of Artificial Intelligence and Regulatory Considerations Let us assume that artificial intelligence is in fact a broad technology revolution not defined by a single capability like self-driving cars. Rather, it is an amalgam of many parts, each moving through the same process at a different rate from different starting points. It can be expected that there will be winners and losers as artificial intelligence advances. Orderly growth requires ethical guidelines governing human behavior in commerce and national defense. The lack of rules governing labor in British factories during the Industrial Revolution, for example, led to unrest, suffering, and ultimately the utopian ideology of Communism. In the spirit of learning from our common history, rather than blindly repeat- ing it, it is now time to consider ethical guidelines for the use of artificial intelligence appli- cations. A practical concern for society is to understand the consequences of the potential misuse of artificial intelligence, which may be addressed through either government regulation or self-imposed rules. The European Union’s General Data Protection Regulation (GDPR 2016/679) is one of the first government-initiated steps to regulate data, which initially focused on ensuring the protection of privacy for individual consumers. Another initiative of the European Union is the EU Regulation on Artificial Intelligence, for which a first pro- posal was presented in April 2021 (Publications Office of the European Union, 2021). The aim of this regulation is to create a common legal framework for the development and use of AI-based systems, which at the same time strengthens the public's trust in these sys- tems by adequately taking into account fundamental rights, security and privacy. Self-regulation with an emphasis on ethics is currently being researched at MIT and Stan- ford University in the United States. Stanford University also recently established the Insti- tute for Human-Centered Artificial Intelligence (HAI) (pronounced “hi”), which is led among others by renowned AI researcher Fei-Fei Li. It serves as an interdisciplinary center combining computer scientists, neuroscientists, and legal scholars focusing on social responsibility. MIT’s College of Computing has similar objectives in promoting the posi- tives of artificial intelligence while preventing negative consequences. 20 Ethical issues are relevant to all commercial and military applications of artificial intelli- gence as well as to the working lives of employees subject to an economy highly influ- enced by artificial intelligence. Relevant considerations include the following: Ethics In the business arena, ethics can be defined in terms of distinguishing between right and wrong conduct and acting accordingly. The choices between the two are many and quan- tified in terms of degree; they are in constant conflict and involve personal and corporate collective consciences. Examples of the unethical use of artificial intelligence include those that support discriminatory behavior towards particular groups of people, take advantage of the disadvantaged, show unchecked zeal for winner-takes-all competition, or that lead to the disbanding of beneficial artificial intelligence applications due to insuf- ficient financial reward. Rights Rights pertain to the physical human body, the right to work, the right to hold property, and to happiness and fellowship. Defining documents for our understanding of rights are the national constitutions or basic laws of states, the United Nations Universal Declaration of Human Rights, and the Geneva Convention, among others. Artificial intelligence projects must observe these rights. In cases where they do not, projects must be changed so that they adhere with them. Governments Governments naturally play an important role in regulating technological advances due to their overarching regulatory and legislative functions. By definition, governments are tasked with the responsibility of preserving a body politic. Artificial intelligence is a sci- ence and, therefore, a common good. This implies that it does not exist without govern- ment control, and it will be regulated sooner or later, as has already occurred in the Euro- pean Union. However, the time lag between industrial innovation and its effective regulation is usually long, and therein lies the danger. Unintended outcomes Unintended outcomes resulting from artificial intelligence technologies pose a significant risk to society, which often result from a lack of both oversight and foresight. The absence of oversight implies a lack of human control once an algorithm has been deployed. Economics Economic considerations are also important. Relevant questions in this domain include the impact of artificial intelligence on the labor market and the unequal distribution of wealth. 21 Artificial Intelligence Viewed as a Revolution When we think of revolutions, violent political overthrows in favor of a new paradigm come to mind. Examples of this are the French Revolution (1789), which involved the top- pling of a monarchy in favor of liberty, equality, and fraternity, the Russian Revolution (1917), which replaced the Czar with Communism, and the American Revolution (1775— 1783), which ended colonialism in favor of a republic. Similarly, humanity has lived through technological and, consequently, economic revolu- tions in the past. Examples include the introduction of the power loom and the spinning jenny (circa 1770), which enabled the global cotton market, steam power, the telegraph and the telephone, the internal combustion engine, sewing machines, and production lines. The artificial intelligence revolution promises to be faster and deeper. Technological revolutions share the overthrow aspect of the term revolution, but they are not necessarily caused by discontent within the masses. Technological revolutions are the result of creative people tinkering and subsequently making discoveries that serve human interests and needs. Gutenberg’s (1440) discovery of the moveable printing process, which led to the broadscale distribution of information, literature, science, scholarship, and the growth of publishing as a business, is a prime example of an innovation that also repre- sented a technological revolution. Other technological revolutions include that brought on by Fritz Haber’s discovery of a workable process for the synthesis of ammonia in 1910 and Ernest Rutherford’s splitting of the atom in 1917. Haber’s discovery led to increases in agricultural yields, with ammonia used as a chemical fertilizer, supporting larger populations and economic growth. By con- trast, Rutherford introduced a new branch of physics which contributed to the develop- ment of the atomic bomb, ending World War II in the Pacific. Spread spectrum (or fre- quency-hopping) which was developed by Hedy Lamarr is also a technology that was originally invented to be used in World War II, i.e. to control torpedoes. Even today, it is still the basis for various technologies in the telecommunications sector, such as Blue- tooth. We are at the beginning of what is publicly described as an artificial intelligence technol- ogy revolution. We do not yet know who the innovators will be. However, we can be cer- tain that collaboration is occurring across the sciences, leading to a brave new world to which we will need to adapt. What the Industrial Revolution and the current revolution in artificial intelligence have in common is that both periods produced inventions that improved people’s lives. The Industrial Revolution used steam and electrical energy to power machinery for mass pro- duction, overcoming the limits of human strength. The artificial intelligence revolution is using computers to produce information-based tools to overcome the information pro- cessing limitations of the human brain. Landmark achievements of the Industrial Revolution include the transition from the steam engine to electrically powered trains and ships, the enhancement of agricultural output as a result of the introduction of mechanized farm machinery and chemical fertilizers, the 22 development of the telephone, and the creation pharmaceuticals, such as penicillin. As a result of the artificial intelligence revolution, we can expect the potential eradication of many common illnesses, more efficient modes of transportation, the development of advanced methods for fighting crime, while simultaneously enabling it with the same technology, and all manner of other discoveries, such as developments in military capabil- ities used for national defense and territorial dominance. It is worthwhile to observe that nearly all newly discovered technologies are based on someone else’s prior discovery, driven by both human needs and innate curiosity. Artificial intelligence is no exception. In ancient times, Finnish people improved the axe, a tool, in terms of its design and the materials used to build it in order to make harvesting trees more efficient. The axe led to better housing and fishing boats, improving people’s diet, ultimately leading to longer life expectancy. The axe is a tool. Artificial intelligence is also a tool that can help the current generation live better and longer. However, just as axes can be used to cause a great deal of damage, so too can artificial intelligence. We can therefore expect that the revolution in artificial intelligence will have both positive and negative consequences for society. 1.2 AI Winters The term “AI winter” was coined around 1984 by artificial intelligence researchers. It is a AI winter bit melodramatic, but it befits the culture of artificial intelligence which is known for its This is a period character- ized by a prolonged exuberance. The word “winter”is borrowed from the expression “nuclear winter”, an after- decrease in interest and effect of a hypothetical nuclear world war, in which ash would overwhelm the atmosphere funding in artificial intelli- and block sunshine from entering the Earth’s atmosphere for years, so nothing would gence research. grow, and temperatures would stay excessively cold. Applied to artificial intelligence, the term refers to a pronounced and prolonged reduction in interest and funding of artificial intelligence technology and research. Such a downturn occurs when it becomes clear that inflated expectations generated in times of exceeding optimism regarding novel techno- logical developments will not realistically be met. There have been two such AI winters, one between 1974—1980 and one between 1987—1993, give or take a few years in either direction. The first AI winter was caused by the non-delivery of promised outcomes in automated language translation and shortcomings in connectionist (network-based) artificial intelli- gence approaches. While the former Soviet Union and the United States were in a cold war, the need for automatic language translation in the West was also great, with not enough translators available to meet demand. Despite initially optimistic assessments to the contrary, early attempts at machine translation failed spectacularly—not in the least due to their inability to handle word ambiguities. Phrases such as “out of sight—out of mind” were translated to “blind idiot”, for example. The United States’ generous govern- ment-funded research was evaluated by the Automatic Language Processing Advisory Committee. It stated that machine translations were less accurate, more expensive, and slower than employing humans. Moreover, it had been shown that the Perceptron—a 23 Perceptron then popular early model of neural-inspired artificial intelligence—had severe shortcom- This is a very simple clas- ings that prevented it from representing even the simple logical function of Exclusive or sification algorithm. In its simplest form, a percep- (XOR). tron is a neural network with a single node. The second AI winter occurred when the artificial intelligence community was overcome by a feeling of general pessimism due to the perception that the industry was out of con- trol, had disappointing results, and would collapse. This perception was based on the demise of the Lisp machine business. Lisp machines were a type of computer that suppor- ted the Lisp language—a popular language for artificial intelligence projects. It became clear, however, that the earliest successful examples of expert systems, which were the main driver of the renewed interest in artificial intelligence during the years 1980—1987, could not be developed beyond a certain point. The reasons for the developmental ceiling were unmanageable growth in the associated fact databases as well as their unreliability with respect to unknown inputs. The counter argument to the notion of AI winters is that these periods of downturn were in fact a myth perpetrated by a few prominent academics, their organizations, and investors who had lost money. Between 1980 and 1990 artificial intelligence was deeply embedded in all kinds of routine processing operations, including credit cards. The fact that Lisp machines collapsed, and expert systems went out of fashion, does not suggest the demise of an entire industry. The Causes of AI Winters AI winters are defined as periods during which funding, research, and interest in the sub- ject significantly decline. The following conditions can contribute to an AI winter: Non-performance: All research is ultimately funded by an interested party, be that a government, a foundation, or by wealthy persons either as an investment or as an act of benevolence. When funded research does not produce promised results, funding even- tually ends. Hype, or the excessive exuberance of promising more than can be delivered, by defini- tion leads to non-performance. Cheaper alternatives than what is being pursued suddenly emerge. This invariably cau- ses sponsors to switch to a new approach, as was the case with Lisp machines. In 1985, the new head of the United States’ Defense Advanced Research Projects Agency (DARPA) only agreed to fund mission-oriented research. Prior to that, DARPA funded basic research projects that by definition were unrelated to any particular mission. Fed- eral Legislation (the Mansfield Amendment of 1969) mandated that DARPA only fund mission-oriented research, and it belatedly complied. Computing capacity: Successful artificial intelligence outcomes often depend on the processing of huge amounts of data, which require a large amount of memory capacity and high processing speed. Availability of data and computing capacity have to be in sync. They have not always been available at the same time. This can lead to the perception of project failure. 24 Actual project failures: When funded research fails, disappointments and cancelled con- tracts are often the result. This was the case with the DARPA funded Speech Under- standing Research (SUR) project in which Carnegie Mellon University failed. However, there is a good argument to be made that basic research will more likely fail than suc- ceed. As a technology, artificial intelligence cannot make big steps until two other technologies are in place: computing capacity in terms of data storage and processing speed and the availability of a large enough volume of data. AI Winter Remedies Research “winters” are not inherently bad, as reform is part of moving forward. This is true for research into artificial intelligence as well as other fields of study. Progress does not look like a hockey stick—that is, flat for a while and then straight up. Remedies for AI win- ters obviously lie in the reduction of factors that bring them about. This is hard to do in a free economy, as no one person or entity controls all these factors. However, one counter- weight to AI winters is competition between nation-states. Competition between nation-states, such as the United States and China today, the United States and the former Soviet Union prior to 1991, and Japan’s Fifth Generation Projects in 1981, has kept opposing nation-states engaged in artificial intelligence research. Engage- ment translates into funding, as artificial intelligence is a dual technology. As early as 1981, Japan’s Ministry of International Trade and Industry (MITI) funded a Fifth Generation Project for software and hardware to perform conversations, translate languages, inter- pret pictures, and perform humanlike reasoning. While obviously not meeting its promise, it got the attention of many governments. It is important to understand that most artificial intelligence technology is dual purpose—that is, artificial intelligence technologies have military value as well as a consumer value. For example, robots and drone aircraft can be used in military battlefields as well as to pick products in warehouses and deliver medical supplies to remote areas. All governments share legitimate interests in defense and the management of their economy in terms of employment and competition. The Next AI Winter The answer to the question of whether there will there be another AI winter is that nobody knows for sure. If we stick with the definition of an AI winter as a period of little or no research funding for artificial intelligence, it is possible that if a hyped concept such as singularity were to get funded, and then defunded as a result of non-performance, an AI Singularity winter could indeed result. However, one argument against AI winters is that many artifi- This refers to the moment when exponential growth cial intelligence technologies wind up being embedded within other fields, such as bioin- in machine intelligence formatics, assuming a different name as a result. The defunding of non-performing accelerates, resulting in projects can, therefore, actually be good because change and renewal are key compo- next-generation artificial intelligence that can use nents of a revolution. The reader is therefore free to decide whether the concept of AI win- its superior resources and ters matters or whether it is merely an interesting myth. capacities for self- improvement to create a subsequent version of ever-improved intelli- gence. 25 1.3 Notable Advances in Artificial Intelligence Having investigated the pronounced downturns in the history of artificial intelligence in the previous section, the focus now shifts to the prosperous times in between previous AI winters, their characteristic research topics, and their successes. Additionally, we will also examine important developments in adjacent fields of study and their relationship to pro- gress in artificial intelligence. 1956—1974: Nascent Artificial Intelligence The early years of artificial intelligence research were dominated by what is now called symbolic artificial intelligence. This approach tries to formalize thought processes as the manipulation of symbolic representations of information according to the rules of formal logic. Consequently, artificial intelligence systems of this era were predominantly con- cerned with the implementation of logical calculus. In this vein, they often implemented a variant of a search strategy in which solutions were arrived at by following a step-by-step procedure in which the steps either followed logically from the preceding state or system- atically tried to explore possible alternatives using backtracking to avoid dead ends. Dur- ing these years, the first attempts at processing natural language were also developed. Since both logical search methods and the first steps towards language processing focused on highly constrained settings and environments, initial successes could be ach- ieved. This so-called microworld approach to creating simplified working environments for the first artificial intelligence solutions also proved to be fruitful in computer vision and robot control. Parallel to these developments, the first theoretical models of neurons—the cells that make up mammalian brains—were developed as corresponding models of how these cells or computational units could interact in networks to implement simple logical functions. 1980—1987: Knowledge Representation While the first wave of artificial intelligence research primarily focused on logical infer- ence, the second wave was predominately characterized by trying to address the problem of knowledge representation. This shift in focus occurred due to the insight that intelligent behavior in day-to-day situations was largely reliant on common sense knowledge about how the world works rather than merely on logical inference. To capture this knowledge- based aspect of intelligence, the technological approach of expert systems was devised. The main feature of this technology was the attempt to systematically store domain rele- vant knowledge in knowledge databases and devise methods to effectively and efficiently navigate them. Moreover, a noticeable upturn in government funding occurred at the beginning of the 1980s, with the Fifth Generation Computer project of the Japanese Gov- ernment and the Alvey project in the United Kingdom being notable examples. Initial set- backs in the development of network and neurally-inspired approaches to artificial intelli- gence were also remedied by newly proposed network models and backpropagation as an effective training method for layered networks of computational units. 26 1993—Today: Learning from Data The 1990s saw some big advances in game artificial intelligence when the first computer system (IBM’s Deep Blue) beat the then world-champion Garry Kasparov. Apart from this substantial but narrow success, artificial intelligence methods have taken hold in engi- neering real-world solutions, and successful approaches in its subfields have gradually entered real-world applications, often without being explicitly referred to as artificial intel- ligence. Moreover, since the early 1990s, there has been an influx of ideas from mathemat- ics, statistics, decision theory, and operations research, which have contributed to artifi- cial intelligence becoming a mature and rigorous scientific discipline. In particular, the intelligent agent paradigm has gained substantial popularity. Here, the notion of the rational agent of economic theory combines with computer science notions of objects and modularity to form the idea of an intelligently acting entity. From this point of view, artifi- cial intelligence can be seen as the study of intelligent agents, thereby freeing it from the notion of imitating human intelligence to a broader study of intelligence in general. These aforementioned advances have been supported by a marked increase in computa- tional and data storage capabilities. This, together with the unprecedented increase in the volume, variety, and velocity of data generation, which occurred during the rise of the internet, has set artificial intelligence up for its current boom. The latest upturn in the pop- ularity of and interest in artificial intelligence research, which began in 2012, is predomi- nantly based on advances in connectionists machine learning models, known as deep learning. In deep learning, theoretical advances in the construction and adaptation of net- worked machine learning models are linked synergistically with the previously mentioned increase in information storage and processing capabilities and the presence of a larger corpora of data to train such models to achieve hitherto unseen levels of performance in many machine learning benchmark problems. In turn, these developments have revived interest in long-established learning models, such as reinforcement learning, paving the way for altogether new ideas, such as adversarial learning. Linguistics Linguistics can be described as the science of natural language in general. It encompasses the study of the structural (grammatical) and phonetic properties of interhuman commu- nication. Understanding language requires information about the subject matter and the context in which it is used. To this end, linguistics distinguishes between the following subfields (McGregor, 2015): Phonetics and phonology: studies speech sounds Morphology: deals with the structure of words Syntax: deals with the structure of sentences Semantics: studies the meaning of words and sentences Pragmatics: studies the use of language Noam Chomsky (1957) was instrumental in establishing this with his book Syntactic Struc- tures. One could go further and consider creativity and thought as related to linguistic arti- ficial intelligence given that all our ways of thinking are so intimately related to language as a form of representation. How, for example, can a child say something it has never said 27 before? In the field of artificial intelligence, we think of natural language as a medium of communication in which the context is a given. Language is, therefore, much more than representation. Cognition In the context of artificial intelligence, cognition refers to several different faculties, includ- ing sensing and perception, learning and understanding, reasoning and intelligence, and thinking and apprehension. The English word “recognition” reflects this. Much of our con- temporary understanding of cognition in artificial intelligence ultimately comes from a combination of several bodies of knowledge. In this case, the fields of psychology and arti- ficial intelligence combine. Psychology deals with experimentation on humans and ani- mals to form new theories and hypotheses; computer science deals with the behavioral modeling of what psychology has observed. Computer models of the brain receive a stim- ulus and make an internal representation of that new stimulus, which can in turn lead the brain to modify its internal representation. If it does, it can lead to actions being taken by the brain. If we want to model the brain, we have to get into it, virtually of course. This can be done by observation or introspection. Once we have a cognitive computer model that works well, the next step is to determine how the model makes decisions, understanding its rationale. It is important for the artificial intelligence industry to provide “explained decisions”, which means that the artificial intelligence model should be able to make its reasoning transparent to an outside observer. As artificial intelligence-based decisions are involved in exceedingly large areas of our lives, the ability to explain how decisions are made is demanded by people concerned with the ethical use of artificial intelligence. Nev- ertheless, this aspect is still lacking, particularly in deep learning-based approaches, which are currently quite popular. Gaming Gaming does not only mean gambling or playing games like tennis. In artificial intelli- gence, it is about learning, uncertainty, and probability. John Von Neuman (1903—1957) and Oskar Morgenstern (1902—1977) established game theory as a formal mathematical field of study. Subsequently, a comprehensive taxonomy of games was established, and in some cases even provably optimal rational gaming strategies were developed. Most games are played for entertainment or the challenge of winning, such as in chess, poker, checkers, and bridge. For all of these games, machines now play at a higher level than even the world’s best human experts. Decision theory and game theory are somewhat related. Decision theory addresses uncer- tainty and usefulness, or stated in another way, probability and utility. Game theory is characterized by the fact that player X’s moves affect player Y’s options. However, winning is not necessarily the objective in each case. Children who play games are not only amus- ing themselves and filling in time; they are also experimenting with options and thereby learning and finding out what works, how things work, with understanding developing from their observations. 28 An example of a zero-sum, perfect information, and yet highly complex game, is the ancient Asian board game known in the West under its Japanese name Go (WeiQi in Chi- nese, Baduk in Korean). Go is a 2,500-year-old two-player game in which the main objec- tive is to surround the most territory. Despite its simplicity in terms of the rule set, its finesse and complexity stem from the large game board that enables a combinatorial explosion of possible moves. Before 2016, it was believed that this combinatorial com- plexity (which prevents the successful application of methods used for games like check- ers and chess) would make it a domain characterized by the dominance of human players for the foreseeable future. However, in 2016 one of the world’s best Go players, the Korean Lee Sedol, played a 5-game match against AlphaGo, a Go-playing system developed by the Google-owned company DeepMind. This match ended in a resounding 4—1 win for AlphaGo—a system based on deep networks and reinforcement learning. Whereas the version used in the 2016 match used Go knowledge acquired by humans dur- ing the long history of the game by employing records from professional Go players for training, DeepMind has subsequently developed a system (named AlphaZero) that learns the game solely based on self-play, foregoing traditional Go knowledge altogether. Remarkably, the resulting system has come out even stronger and has subsequently played numerous games where it has surprised human experts with effective and efficient moves that were hitherto shunned by orthodox Go theory. Such an approach is particu- larly useful in policy research and the discovery of strategy options humans have not yet come up with. The Internet of Things (IoT) The Internet of Things (IoT) refers to interconnected computing devices that are embed- ded in homes, cars, on wearable devices, and phones. Such devices are typically used every day. In addition, such devices are enabled to send and receive data amongst them- selves and their users. This generates large amounts of data. The possibilities offered by AI are ideal for evaluating this data and providing further services based on it. It is important to remember that this data can be used for both productive and destructive purposes. A constructive application may take the form of medical reminders to take medication based on the physical measurements generated by wearable devices. A negative applica- tion could be the monitoring of data about the medications a person is taking in order to determine their health insurance rate. In other words, privacy violations and the ethics of data use are issues facing the use of artificial intelligence devices. While IoT does not qual- ify as artificial intelligence per se, it is linked to it. This is because IoT is about connecting machines and generating data, and the intelligent behavior of these machines is derived from artificial intelligence algorithms. Quantum Computing Quantum computing, like the Internet of Things, is a technological advancement that, when combined with artificial intelligence, enables innovation. The term “quantum com- puting” comes from the physical theory called quantum mechanics which deals with the strange behavior of sub-atomic particles. An electron, for example, can be in two differ- ence states (places) at the same time. In general, quantum mechanics posits that physical systems are characterized by a wave function describing the probabilities of the system in 29 any particular state. The goal of quantum computing is to build super computers that uti- lize these quantum properties of matter to implement novel algorithmic approaches that significantly outperform classical machines (see Giles, 2019). It can be reasoned that the probabilistic approach inherent in many artificial intelligence techniques and the ubiqui- tous use of optimization methods are well suited to the kind of information processing that quantum computers provide, putting it in a very promising position to profit from new technological advances in the field. SUMMARY This unit provided an initial overview of the field of artificial intelligence, examining notable discoveries that were instrumental in the develop- ment of the field, such as logical calculus, branch and search algorithms, knowledge representation in expert systems, and networked models of machine learning. It also took a look at failures and periods of stagna- tion along the way. The goal has been to demonstrate that during the historical develop- ment of artificial intelligence, different paradigms have at different times shaped public understanding of the field. Consequently, there is no hard and fast reason to believe that the currently popular paradigm of deep learning networks offers the final word on how to implement intelligent behavior into machines. Additionally, some adjacent fields to artificial intelligence, such as lin- guistics, gaming, the Internet of Things (IoT), and quantum computing have been touched upon to give a flavor of the broad array of fields that have either contributed to the development of artificial intelligence or have been influenced by it. 30 UNIT 2 EARLY SYSTEMS IN ARTIFICIAL INTELLIGENCE STUDY GOALS On completion of this unit, you will have learned … – about important approaches that have defined the field of artificial intelligence in the past and that continue to influence it today. – why expert systems are important and how they have contributed to artificial intelli- gence and computer science. – about advances brought about in the Prolog programming language. – the definition of machine learning and how it contributes to artificial intelligence. 2. EARLY SYSTEMS IN ARTIFICIAL INTELLIGENCE Introduction Throughout the history of artificial intelligence, a diverse set of approaches have been explored to tackle the problem of emulating cognitive processes and capabilities. Some of these have been virtually abandoned by the scientific community while others are still being actively pursued to this day. However, most of them have experienced great varia- bility in popularity over the course of the last 70 years of research in artificial intelligence. Notably, even abandoned branches of artificial intelligence have brought forth valuable insights into different aspects of intelligence, highlighting the intricacy of cognitive pro- cesses and dispelling many early misconceptions about the supposed simplicity of per- ception and cognition-related tasks. In this unit, three important branches of artificial intelligence research are introduced, each representing a major vantage point of artificial intelligence, significantly advancing perception of research in the field. 2.1 Overview of Expert Systems As the name suggests, the goal of expert systems is to emulate the decision and solution finding process of an expert. The word “expert” refers to a human being with specialized knowledge and experience in a given field, such as medicine or mechanics. Since prob- lems in any given domain may be similar to each other, but never quite alike, solving prob- lems in a given domain cannot be accomplished by memorization alone. Rather, problem- solving is supplemented by a method matching experiential knowledge to new problems and application scenarios. Expert systems are therefore composed of a body of formalized knowledge and an inference engine that uses the knowledge base to draw conclusions. With respect to the representation of knowledge, three main approaches to expert sys- tems can be distinguished: Case-based systems store examples of concrete problems together with a successful solution. When presented with a novel, previously unseen case, the system tries to retrieve a solution to a similar case and apply this solution to the case at hand. The key challenge is to define a suitable similarity measure to compare problem settings. Rule-based systems represent the knowledge base in the form of facts and if-A-then-B- type rules that describe relations between facts. If the problem class to be solved can be categorized as a decision problem, the knowl- edge can be represented in the form of decision trees. The latter are typically gener- ated by analyzing a set of examples. 32 The inference engine, on the other hand, implements rules of logical reasoning to derive Decision tree new facts, rules, and conclusions not explicitly contained in the given corpus of the knowl- A visual representation of multi-decision processes edge base. in the form of a tree dia- gram is called a decision Historically, expert systems are an outgrowth of earlier attempts at implementing a so- tree. Decision alternatives can be read from the called general problem solver. This approach is primarily associated with the researchers branches with their Herbert A. Simon and Allen Newell, who used a combination of insights from cognitive sci- respective ramifications (nodes). ence and mathematical models of formal reasoning to build a system intended to solve arbitrary problems by successive reduction to simpler problems in the late 1950s. While this attempt ultimately has to be considered a failure when compared to its lofty goals, it has nevertheless proven highly influential in the development of cognitive science. One of the initial insights gained from the attempt at general problem solving was that the construction of a domain specific problem solver should—at least in principle—be easier to achieve. This led the way to thinking about systems that combined domain specific knowledge with domain dependent apposite reasoning patterns. Edward Feigenbaum, who worked at Stanford University, the leading academic institution for the subject at the time, defined the term expert system and built the first practical examples while leading the Heuristic Programming Project. The first notable application was DENDRAL, a system for identifying organic molecules. Given data and rules, the next step was to establish expert systems to help with medical diagnoses of infectious diseases. The expert system that evolved out of this was called MYCIN, which had a knowledge base of around 600 rules. However, it took until the 1980s for expert systems to reach the height of research interest, leading to the development of commercial applications. The main achievement of expert systems was their role in pioneering the idea of a formal, yet accessible representation of knowledge. This representation was explicit in the sense that it was formulated as a set of facts and rules that were suitable for creation, inspec- tion, and review by a domain expert. This approach thus clearly separates domain specific business logic from the general logic needed to run the program—the latter encapsulated in the inference engine. In stark contrast, more conventional programming approaches implicitly represent both internal control and business logic in the form of a program code that is hard to read and understand by people who are not IT experts. At least in principle, the approach championed by expert systems enabled even non-programmers to develop, improve, and maintain a software solution. Moreover, it introduced the idea of rapid pro- totyping since the fixed inference engine enabled the creation of programs for entirely Rapid prototyping different purposes simply by changing the set of underlying rules in the knowledge base. This is a procedure in which prototypes are built and evaluated as However, a major downside of the classical expert system paradigm, which also finally led quickly as possible. to a sharp decline in its popularity, was also related to the knowledge base. As expert sys- Through rapid prototyp- ing, feedback on impor- tems were engineered for a growing number of applications, many interesting use cases tant functionalities can be required larger and larger knowledge bases in order to satisfactorily represent the domain obtained from prospec- tive users early in the in question. This insight proved problematic in two different respects. Firstly, the compu- development process tational complexity of inference grows faster than it does linearly in the number of facts and rules. This means that for many practical problems the system’s answering times were 33 Consistency prohibitively high. Secondly, as a knowledge base grows, proving its consistency by This refers to a set of logi- ensuring that no constituent parts contradict each other, becomes exceedingly challeng- cal propositions free of contradictions in which ing. all propositions are true at the same time. A set of The construction of inference engines for expert systems highlighted the need for a pro- propositions in which all statements cannot be gramming language that facilitated the formulation of logical rules and reasoning pro- true at the same time is cesses. To this end, the programming language Prolog, meaning “programming in logic” called inconsistent. or “programmation en logique” in French, became relevant. 2.2 Introduction to Prolog Prolog was created by French computer scientists Alain Colmerauer and Philippe Roussel, with the logician Robert Kowalski further developing the language. It was first implemen- ted in the early 1970s. The main motivation for creating Prolog was to use it in the devel- opment of systems for natural language processing and artificial intelligence. The aim of this section is not to gain programming proficiency in Prolog, but rather to gain an appre- ciation of the language as a tool for solving logical problems and of its contribution to the development of artificial intelligence and the design of programming languages. At the most basic level, a digital computer processes information in the form of 0—1 values designated as bits. Clearly, this form of representation is not ideally suited for human interpretation and manipulation. In order to facilitate the programming of such a device, programming languages have been designed to provide abstractions to the fundamental technical layer that are closer to human thinking, algorithmic description, and reasoning patterns. The most important difference between programming languages stems from the degree and kind of abstractions that are introduced. Most of the computer languages developed during the period in which Prolog was conceived and implemented were imperative languages—that is, languages encoded in a program as a series of instructions for the machine to follow in order to produce a desired outcome or solution. Declarative In contrast, Prolog is based on a declarative programming paradigm. The programmer programming specifies characteristics of the desired solution and the programming language interpreter This is a programming style in which the pro- then constructs a sequence of processing steps to reach the given goal. A prominent exam- grammer specifies the ple of this paradigm is the structured query language (SQL) for relational databases. A typ- properties of the sought ical query is given by a statement specifying the table from which records are to be solution but not the algo- rithm—that is, the retrieved together with one or more conditions the records should fulfill. The database sequence of operations management system then automatically generates an execution plan—a sequence of pro- that lead to a solution. cessing steps—that produces the outcome as specified by the query. Analogously, a Prolog program consists of a collection of facts and rules that relate the facts to one another. Pro- gram execution is then initiated by formulating a query using the aforementioned knowl- edge base. Before we dive more deeply into Prolog, consider the following analogy: 34 1. As a human being, you have a brain that is full of data, facts, numbers, and bits and pieces of knowledge accumulated throughout the course of your life. Think of this as your knowledge base. 2. You also understand rules that go with facts, many of which you have observed and applied over time. Think of these as logic rules that when applied result in good deci- sions. 3. You are also curious and want to learn about changing your surroundings for the bet- ter, and so you often ask questions. To come up with answers, you draw on the facts, apply the rules, and get reasonable, common sense answers that hopefully constitute solutions to perceived problems. Prolog is designed to formalize these processes in the form of first order logic. Prolog’s First order logic structure is made up of predicates and clauses. A predicate is a Boolean function that This is a branch of mathe- matical logic, which is assigns a truth value to some object X. As such, predicates are commonly used to describe also called predicate properties of objects. The term “clause” denotes a logical expression formed from a finite logic. number of literals. Prolog programs typically start by declaring facts and relationships. For example: A and B are both male. A and B have the same father. A and B have the same mother. A and B are not the same. Another relationship declaration could be between a person and a piece of property. For example, the statement “Joachim owns a book” declares a relationship of ownership between Joachim and the book. Once basic relationships are declared, facts can be considered, questions can be asked, variables can be included, goals can be formulated, and patterns can be matched. What follows is a very small selection of very basic statements illustrating the nature of the lan- guage. Statements are always in lower case letters and variables start with a capital letter. Table 1: Example of the Prolog Language Prolog language construct Prolog syntax Meaning and output Fact lectures (Smith, DLMAIAI01) Establishes the fact that Dr. Smith teaches the course DLMAIAI01. It is an example of a Prolog clause. Predicate professor/1 Defines the one argument predi- professor(Smith). cate professor by three facts. professor(Jones). Drs. Smith, Jones, and Meyer are professor(Meyer). professors. Rule technicalCourse(X) :– engineer- All engineering courses are tech- ingCourse(X) nical courses. Note the use of variable X! 35 Prolog language construct Prolog syntax Meaning and output Query ? – lectures(Smith, DLMAIAI01) Does Dr. Smith teach DLMAIAI01? Goal ? – lectures(Smith, X) What courses does Dr. Smith teach? Note the use of the varia- ble X! Source: Created on behalf of IU (2019). As Prolog was uniquely well adapted to handling logic, and querying knowledge bases, it became instrumental in a variety of commercial applications (Roth, 2002), as listed below: Environmental studies modeling weather phenomena were conducted at Penn State University, with Prolog used to build a system for weather forecasting and air pollution dispersion. The University of Surrey in the United Kingdom developed several systems for water utilities, with Prolog used for water distribution and planning, especially in cases of emergency. Manufacturing always seeks to reduce costs. Prolog gained some recognition by Boeing, the aircraft manufacturer, for its development of a system called CASEy, which directs shop floor workers in the application of electrical parts and in how to follow proper operational procedures. This led to a reduction in assembly times. 2.3 Pattern Recognition and Machine Learning (ML) The field of machine learning is as old as artificial intelligence itself. However, it only recently became the dominant paradigm in artificial intelligence research. One of the most often cited operational definitions of machine learning was coined by the American researcher Tom Mitchell (1997, p. 2): “A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P if its performance at tasks in T, as measured by P, improves with experience E.” This definition underlines the fact that learning from data is a key characteristic of machine learning. To this end, it draws upon a multitude of methods from classical statistics to more algorithmically moti- vated approaches. In order to gain a better overview of machine learning it is helpful to distinguish between some prominent approaches to learning. Basically, the following three types can be distin- guished (Russell & Norvig, 2022; Marsland, 2014; Murphy, 2012): Supervised learning Supervised learning operates on labeled data sets, meaning that the learning examples consist of object descriptions in the form of features, together with labels pertaining to these objects. Tasks can then be described using the given examples to identify mapping 36 between feature values and outputs that enable the learner to predict the label for hith- erto unseen objects. Depending on the kind of output that is to be produced, one distin- guishes between regression and classification. In regression, the output is a continuous numerical variable. Thus, regression aims at finding real-value functions that represent mapping between the input space of features and the output space of associated values. On the other hand, if the output is restricted to a limited set of values, one speaks of classi- fication. Common examples include labeling e-mail messages as spam or finding images of certain content in large image databases. Unsupervised learning Unsupervised learning operates on data without any labeling information. The primary goal is to identify structures or patterns in the data. The most prominent examples of unsupervised learning techniques include clustering,(ie. finding groups of data points with high similarity), dimensionality reduction techniques (ie. constructing low-dimen- sional projections of potentially high dimensional feature spaces that, at the same time, preserve an interesting structure), and statistical techniques that estimate the probability density functions of random variables. Reinforcement learning Reinforcement learning considers a learning agent in an environment. The agent can per- Agent form actions that influence its internal state and that of the environment. A reward func- In the field of artificial intelligence, the term tion is employed to judge the utility of the performed actions with respect to a stated goal. agent refers to an autono- Since the agent creates its own learning data through trial and error testing of action alter- mous entity that per- natives, no prior data collection is necessary. Due to the setting of the learning problem, ceives its environment and acts on it in a goal- reinforcement learning is often associated with, or guided by, results from game and deci- oriented manner. sion theory. A prominent example of an artificial intelligence system that employs rein- forcement learning techniques is AlphaZero. Examples of this system learned to play the games Go, chess, and shogi at superhuman playing ability by only using knowledge of the basic rules and extensive self-play. 2.4 Use Cases In this unit, three major currents in the development of artificial intelligence have been identified. The following list provides a brief but by no means an exhaustive overview of the multitude of problems that are being addressed using artificial intelligence. Health care ◦ Wearable devices, not unlike a wristwatch, can monitor critical signs of life, such as blood pressure and body temperature. From this data, an artificial intelligence agent or expert system can dispense advice relative to the conditions of the wearer. ◦ Given multiple medical conditions, a prescription agent can suggest treatment options in terms of the optimal combination of prescriptions in order to avoid nega- tive side-effects. 37 ◦ An artificial intelligence agent can be used to monitor a physician’s patients and their respective needs to ensure that appointment deadlines are met, especially when there are many patients. Automobiles and transportation ◦ While automobiles are not yet fully autonomous in normal traffic, the average car is equipped with numerous sensing devices to assist the driver in remaining safe. ◦ Artificial intelligence sensors can detect technical problems originating from the car as well as medical conditions emanating from the driver, such as a driver’s alcoholic breath. Banking ◦ Major fraud has been detected years after it occurred. Even minor day-to-day irregu- larities are detectable with pattern recognition technologies using artificial intelli- gence. ◦ Counterfeit signatures are more easily detected due to the scanning of originals into a database. ◦ Robo-advising is also now offered by banks and broker-dealers. Based on a rich vari- ety of securities and an investor’s risk profile, a robo-advisor can be used to construct an appropriate portfolio. Manufacturing ◦ Certainty in correctly assembling >3000 aircraft parts can now be achieved with expert systems. ◦ Artificial intelligence technology is good at evaluating all of the many different possi- bilities facing the design of new products. It can, therefore, be used to assist in the creative design process. Education ◦ In online instruction, personalization can significantly improve the quality of teach- ing. This is particularly important when individual support is difficult due to the high number of participants, such as in Massive Open Online Courses (MOOCs). ◦ The timely grading of test results both quantitatively in terms of numerical grades and qualitatively in terms of verbal responses can be enhanced with artificial intelli- gence technologies. Retail ◦ Websites can track how interests change based on the number of website visits and purchases made. In cases where website visitors can be identified, artificial intelli- gence can make personalized purchase predictions. ◦ Chatbots of the future will know callers via voice recognition and display patience, good humor, good manners, and even be kind. To this end, chatbots support cus- tomer retention and customer service. ◦ Market segmentation used to be based on geographical regions, such as state, prov- ince, or county. It is now possible on a street-by-street basis. The options are endless and personal. For example, your vacuum cleaner will have learned your living room layout without missing a corner, minimizing the distance trav- eled, and your personal instantaneous translator will help you communicate with the locals on your next overseas holiday. 38 The examples above briefly summarize applications in which artificial intelligence will probably play a significant role. The following case study illustrates just how this technol- ogy will assist in meeting urgent company needs. BUSINESS CASE STUDY This case study involves the Mizuho Bank in Japan, which is headquartered in Tokyo and has over 500 branches. The issue at hand was customer service inter- action, which is varied and complex in the banking industry. It is not like a ware- house where all one has to do is consult an inventory list and answer customer questions with “yes we have that part, and the price is…”. Banking questions often involve international transfers, local regulations, tax issues, fraud, invest- ment advice, and interest lending rates. Mizuho’s ambition was to analyze customer conversations in real time using a natural language processing (NLP) algorithm so that employees answering cus- tomer enquiries had the best information available on their computer screens, enabling them to give good, real-time responses. This case illustrates how artifi- cial intelligence technology can be used as a tool to help humans at work. The bank’s objective was to improve employee performance in responding to customer calls, especially those of new employees with less practical experi- ence. The bank’s methods included the “Cloud”, internet, NPL algorithms, statis- tics, and continuous learning as a result of the algorithm listening to phone calls. The results included: 1. A higher level of customer service 2. A reduction in employee response time to customer questions 3. A reduction in call center staff quality improvement training However, assume that customer conversations were recorded without a cus- tomer’s permission. Do you think this constitutes an ethical violation? If so, why? If not, why not? SUMMARY This unit has highlighted three main currents in artificial intelligence research that have shaped the field at various times throughout its his- tory. Expert systems try to emulate the knowledge and decision-making capabilities of human experts. To this end, facts and their governing rules from a certain domain are encoded in a machine-readable form in 39 a knowledge base. An inference engine operates on that knowledge base in order to derive new and hitherto unknown facts and relations that can be used to make decisions or solve problems in the pertaining domain. Prolog was introduced as the primary example of logic programming. Logic programming tries to implement first order logical reasoning. Facts, rules, and relations are formulated via predicates and clauses. Programming is done in a declarative way by querying the knowledge base of facts and rules without explicitly specifying the steps that lead to a solution. As with artificial intelligence, the scientific field of machine learning was established in the late 1950s. Machine learning employes algorithmically motivated techniques and approaches that draw upon statistics to learn from data. Depending on the kind of data used, different types of learn- ing can be distinguished. While supervised learning depends on labels, unsupervised learning is more concerned with the identification of structures and regularities in data. Reinforcement learning is based on the concept of an agent that explores its environment through actions that lead to a reward based on their

Use Quizgecko on...
Browser
Browser