CMU-Q 15-381
Lecture 1: Introduction – AI, basic definitions, problems, road map
Teacher: Gianni A. Di Caro
CMU-Q 15-381 Lecture 1: Introduction AI, basic definitions, - - PowerPoint PPT Presentation
CMU-Q 15-381 Lecture 1: Introduction AI, basic definitions, problems, road map Teacher: Gianni A. Di Caro O UTLINE AI? Representation and Problem Solving? What this course is about Fundamental definitions and notions Road map
Teacher: Gianni A. Di Caro
2
3
In fiction, popular views …
4
In the real world …
2016 Libratus wins $1.7m in chips at Texas Hold’em AI system from CMU has beaten four of the world’s best poker players in a 20-day tournament. 2016-2017 AlphaGo beats world’s top Go players AI system from DeepMind /Google CMU has beaten Go’s grandmasters in 3-games matches
5
6
In the real world …
2014 Alexa: virtual personal assistant, smarthome 2016 J. Bezos’ WP news writer, since Rio Olympics
7
In the robot world …
8
In the daily digital life …
9
The science of making machines do things that would require intelligence if done by man (Bertram Raphael) Many different views, but all share a common concept … Intelligence ?
10
(Intelligence) The cognitive ability of an individual (entity) to learn from experience, to reason well, to remember important information, and to (effectively) cope with the demands of daily living (Robert Sternberg) Artificial intelligence is that activity devoted to making machines intelligent, and intelligence is that quality that enables an entity to function appropriately and with foresight in its environment (Nils J. Nilsson, The Quest for Artificial Intelligence, 2009)
11
Artificial intelligence is that activity devoted to making machines intelligent, and intelligence is that quality that enables an entity to function appropriately and with foresight in its environment (Nils J. Nilsson)
Is quicksort an example of AI? Is a pocket calculator an example of AI? q Scale q Autonomy q Speed q Flexibility q Generality q Capability q …. Intelligence ~ Human-related notion
§ Participants included Marvin Minsky, John McCarthy, Claude Shannon, Ray Solomonoff, Arthur Samuel, Allen Newell, Herbert Simon
§ “We propose that a two-month, ten-man study of artificial intelligence be carried out during the summer of 1956 at Dartmouth College in Hanover, NH. The study is to proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it.”
12
13
The science of making machines that:
Think like people Act like people Think rationally Act rationally
Thought processes and reasoning Behavior Fidelity to human performance Ideal performance Functionally equivalent, imitation: it doesn’t matter how. Turing Test Theory of the mind ⟷ AI computer models, Cognitive science An agent that does the “right thing” based
doesn’t matter). Rational agents Laws of thought to represent and reason about things: Logics 973537498401 ?
– Weak AI: An artificial intelligence system can (only) act like it thinks and has a mind
14
Philosopher J. Searl (1980) made a distinction between different hypothesis about AI Strong AI: An artificial intelligence system can think and have a mind
R&N, Ch. 26
– Narrow AI: An artificial intelligence system that replicates and maybe surpasses human intelligence for a dedicated purpose. “Applied” AI to a “narrow” domain.
15
General AI: A general-purpose artificial intelligence system that replicates and maybe surpasses human intelligence. A system with comprehensive knowledge and cognitive capabilities. Processing speed and capacity can be larger than humans. Embodiment (?) → Strong AI
Elon Musk: AI is “our greatest existential threat.” Stephen Hawking: “Success in creating AI would be the biggest event in human
last...” Bill Gates: “First, the machines will do a lot
That should be positive if we manage it
the intelligence is strong enough to be a concern.”
16
Emergence of superhuman intelligence Key idea: self-improvement Source of name: analogy between inability to predict events after the development of a superintelligence, and the space-time singularity beyond the event horizon of a black hole Some predict: this century Others argue: never
17
18
19
20
§ Data, data, data… to learn from!
§ Fast computers § GPUs § New techniques http://www.nature.com/nature/journal/v518/n7540/abs/nature14236.html
21
Think like people Act like people Think rationally Act rationally
Thought processes and reasoning Behavior Fidelity to human performance Ideal performance Functionally equivalent, imitation: it doesn’t matter how. Turing Test Theory of the mind ⟷ AI computer models, Cognitive science An agent that does the “right thing” based
doesn’t matter). Rational agents Laws of thought to represent and reason about things: Logics
22
This course: “Solving problems in the best possible way!” = Computational Rationality + (Modeling + Problem Solving)
= Mathematical and computational techniques for rational decision-making ↔ Act rationally
§ Provably optimal § Provably sub-optimal § Heuristic
Fundamental question: How to issue the sequence (≥ 1) of decisions that best/optimally (given accessible information and data) achieve my performance objectives?
23
… a single decision, or a sequence of decisions ...
24
Not to be absolutely certain is, I think,
(Bertrand Russel) Definitions are the guardians of rationality, the first line of defense against the chaos of mental disintegration (Ayn Rand) In everything, one thing is impossible: rationality (Friedrich Nietzsche)
25
1. Representation of the Problems / World Modeling!!! 2. Declare Goals and Preferences → Performance measure: Objective criterion to asses degree of success
available information (from sensors+built-in)
4. Make decisions (act) that, given (1+2+3), maximize the expected performance (with formal guarantees)
More things should not be used than are necessary (Ockham)
Rationality ≠ Omniscience
26
1. How to make effective abstractions in order to categorize problems and define efficient problem representations 2. How to define goals, preferences, and utilities to effectively direct the problem solving process
Problem Formal Model Abstraction Problem Class My goal is to get to West-Bay (possibly in the minimum time) I don’t have a specific goal, but different states have a different utility (numeric value) I have different preferences
27
learn, optimal decisions: defining how to act
Closed loop control State A State B Action a
Policy: (state → action) → maximize utility
Open loop control
Plan: (action a, action b, action c, … action n) → goal Classification: pattern → class Plan: (𝑦, = 𝑏, 𝑦/ = 𝑐, 𝑦1 = 𝑑, … 𝑦3 = 𝑜) → goal Regression: observed input → predicted output
Learning a discrete mapping Learning a function
28
Agent Function, for a Rational Agent: For each possible input, the agent selects an action (1) that is expected to maximize the performance measure (2), given the evidence provided by the percept sequence (3) if need, and whatever built-in knowledge the agent has (4)
Agent Sensors Actuators Environment Percepts Actions
?
States of the representation: feasible configurations expressed in terms of selected variables of interest
§ Rational: maximally achieving pre-defined goals (based on representation, accessible data and information) § Rationality only concerns what decisions are made (not the thought process behind them)
29
§ Elements: Vacuum-cleaner robot, dirt, two (interconnected) rooms § Actions: Left, Right, Suck, NoOp § World States: 8 feasible configurations
30
§ Actions: Left, Right, Suck, NoOp § Percepts: Location and Status, e.g., [A, Dirty] function Vacuum-Agent ([location, status]) returns an action
Is the agent behavior rational?
31
function Vacuum-Agent ([location, status]) returns an action
§ Environment: Static § Goal: Suck all the dirt in the minimum time § Environment: Static § Goal: Maximize collected reward {+1 dirt cleaning, -1 moving} § Environment: Dynamic § Goal: Maximize collected reward {+1 dirt cleaning, -1 moving} § Environment: Static, dirt is placed at the second time step § Location sensor: Unreliable, except for the first location § Goal: Suck all the dirt in the minimum time
32
Agent Sensors Actuators Environment Percepts Actions
? Agent behavior: Agent function that maps percept sequences to an action (we will learn how to design it for a rational agent!) Task environment (~Problem): Environment + Actuators + Sensors + Performance A vast range of task environments arise in AI …
33
Chess with clock Taxi (self-)driving
34
Ø Fully observable vs. Partially observable
aspects that are relevant (necessary) to the choice of an action in relation to the selected performance measure Ø Deterministic vs. Stochastic
state, then the environment is deterministic.
stochastic
deterministic environments Ø Known vs. Unknown
probabilities) of an action are given (the agent knows the “rules”)
35
Ø Episodic vs. Sequential
and episodes are independent from each other.
decisions (the agent needs to think ahead) Ø Static vs. Dynamic
Ø Discrete vs. Continuous
Ø Single agent vs. Multi-agent
the impact they have on each other’s performance
36
Different task environments require different techniques for problem representation, agent design, and rational decision-making
37
Different task environments require different techniques for problem representation, agent design, and rational decision-making
38
Problem Representation: states, transitions Constraint Satisfaction models Policy-based sequential decision-making in stochastic environments Sequential action learning through direct interaction with stochastic unknown environments Games: Decision-making in multi-agent adversarial environments Deep learning architectures Continuous and discrete
for decision-making Automated planning in fully observable and deterministic environments Supervised learning of regression models and classifiers Modeling and prediction of Markov random processes Probabilistic inference
39
40
41
By the end of the course, you should have learned a number of different techniques for decision-making and learning, and you should be able to: 1. Identify the type of an AI problem. 2. Formulate the problem as a particular type. 3. Identify the most appropriate technique to use for decision- making. 4. Compare the difficulty of different versions of AI problems, in terms of computational complexity and the efficiency of existing algorithms. 5. Implement, evaluate and compare the performance of various AI algorithms. Evaluation could include empirical demonstrations and/or theoretical proofs.
42
§ Based on lectures § Reference to textbooks and additional material will be provided topic by topic, however, everything you need will/should be covered by lecture handouts § 10 homework assignments: theory and programming (in Python + use of the Anki Cozmo robot) § One midterm + One final § Grading: 35% Final, 15% Midterm, 50% Homework § No smartphones or laptops (when not used for taking notes)
43
Piazza: used for questions, announcements, and polls My office: office hours + drop in at any time! (or send an email)