ILLUMINATING AI: UNDERSTANDING AI'S GOALS, REASONING & - - PowerPoint PPT Presentation

illuminating ai
SMART_READER_LITE
LIVE PREVIEW

ILLUMINATING AI: UNDERSTANDING AI'S GOALS, REASONING & - - PowerPoint PPT Presentation

ILLUMINATING AI: UNDERSTANDING AI'S GOALS, REASONING & COMPROMISES AI says 7 (99%) Image Tsvi Achler MD PhD Tsvi Achler MD PhD If SIRI makes a mistake, the impact is limited In most applications a mistake has more serious


slide-1
SLIDE 1

ILLUMINATING AI:

UNDERSTANDING AI'S GOALS, REASONING & COMPROMISES

AI says “7”

(99%)

Tsvi Achler MD PhD Tsvi Achler MD PhD Image

slide-2
SLIDE 2

If SIRI makes a mistake, the impact is limited

AI Adoption Requires Transparency: Trust, Regulation, and Understanding of the AI’s Compromises But the problem is: AI is a Black Box

2

Banking | Medicine | Self-Driving Cars

In most applications a mistake has more serious consequences:

slide-3
SLIDE 3

With undetectable noise Google

+ =

The Lack of Transparency Leaves You Asking: What Is the AI Really Recognizing? DARPA commitment: “Explainable AI” Initiative

3

EU Legislates: Users Have a Right to an Explanation

Before

slide-4
SLIDE 4

Solution Pathways for Explanability (1) DARPA:

Trial & Error to Find What Effects the Network Mirrors the brain’s ability to provide reasons

(2) Optimizing Mind:

Takes Time & the AI Remains a Black Box The Ground Truth Of The AI

slide-5
SLIDE 5

5

slide-6
SLIDE 6

The Brain Relies on Feedback

During Recognition

AI Does Not

Feedback: No Feedback:

slide-7
SLIDE 7

Feedback is Found Throughout the Brain

(eg Aroniadou-Anderjaska et al 2000)

Tri-synaptic connections

More Feedback Than Feedforward

Retrograde Signaling e.g.: nitric oxide

slide-8
SLIDE 8

AI Exclusively Uses Feedforward Connections W During Recognition

Outputs Y Inputs X

Feedforward Caricature Computational Caricature Mathematical Function

W=

Outputs Y Inputs X

Connectivity Notation Weight Matrix

Y X

W

X W Y   

(during recognition)

Recognition

8

Do not be mislead:

Even when an AI is called

“recurrent” it still uses W

slide-9
SLIDE 9

Why Is Feedback Needed? For Optimization

  • What is Optimization?
  • Difference between “Feedforward” and

Feedforward-Feedback methods

  • Why lack of Feedback During Recognition matters

9

slide-10
SLIDE 10

Optimization

Try configuration, evaluate, modify and repeat until optimal fit

Example: Solving Jigsaw Puzzle

(OP)

OP

Try Evaluate Modify

10

slide-11
SLIDE 11

OP

Recognition Algorithms:

Y X

Bird Bird

Bird

Recall Recognition- Inference Learning Memory Learned feedforward weights in: Deep, Convolutional, Recurrent, LSTM, Reinforcement Networks, Support Vector Machines (SVM), Perceptrons, “Neural Networks” … everything learned via Backprop etc.

by Optimizing (OP) during learning

W

Fodor & Pylyshyn (1988) Sun (2002)

WX Y 

OP

Optimize weights so that recognition occurs using a simple multiplication 1) Optimized weights W are a “Black Box “Feedforward” methods

slide-12
SLIDE 12

OP

Recognition Algorithms:

Y X

Bird Bird

Bird

Recall Recognition- Inference Learning Memory

W

Fodor & Pylyshyn (1988) Sun (2002)

WX Y 

OP

encodes uniqueness into weights “Feedforward” methods

slide-13
SLIDE 13

Determining Uniqueness is essential to perform efficient recognition Problem: Uniqueness changes with context cannot learn uniqueness for all possible contexts

O O O O O X X X O X O

Besides: relevant context is during recognition, not learning

Unique thus Important! Unique thus Important! Training Instance 1 Training Instance 2

slide-14
SLIDE 14

OP

We suggest uniqueness is determined during recognition instead

Bird Bird

Bird

Recall Recognition Learning Memory

2) Optimizing only current test pattern Weights “Clear Box” 1) Determining activation Y (not weights) → Reducing computational costs OP (not all of training data) By Optimizing (OP) during recognition

when the context is available

while estimating uniqueness

slide-15
SLIDE 15

15

slide-16
SLIDE 16

Model-type During Learning (weight Δ) During Recognition (find Y)

Why would the brain only use feedback during learning?

Optimization “Feedforward” Feedforward recognition Simpler Learning M Illuminated AI to find weights W Switch Dynamics to find neuron activation Optimization

Illuminated AI Switches When Optimization Occurs

Requires feedback for learning: for example to back-propagate error Not really Feedforward !!!

slide-17
SLIDE 17

Recognition with Illuminated AI

Outputs Y Inputs X

Symmetrical Inhibitory connections modulate inputs using output activity

M=

Outputs Y Inputs X

Y X

M

M

For optimization Weights are Expectations (allows explainability & update)

Neuron Caricature Computational Caricature Connectivity Notation

17

slide-18
SLIDE 18

Same Results But Transparent Method

X W Y   

pattern from the environment (input) resulting neuron activation (output) "feedforward" weights

SAME SAME

M

Illuminated Networks:

Y  X 

OP

Feedforward:

SAME

Y  X  Y 

18

Easier to Explain, Learn and Update Illuminated Weights

slide-19
SLIDE 19

Example: You train your AI It gets good grades (performance) You are done … right?

Learn Digits

95%

19

slide-20
SLIDE 20

1) Can convert existing feedforward networks to xRFN & see what they are doing: Black Box -> “Clear Box”

MNIST Demonstration

SVM Feedforward xRFN

See Inside!

Equivalent

Overall Accuracy 91.65% By Digit: 1 2 3 4 5 6 7 8 9 False Positives 45 67 106 79 86 67 75 152 109 49 False Negatives 19 129 91 71 137 50 83 125 109 21 SVM Overall Accuracy 91.65% By Digit: 1 2 3 4 5 6 7 8 9 False Positives 45 67 106 79 86 67 75 152 109 49 False Negatives 19 129 91 71 137 50 83 125 109 21 RFN

Why are Explainable Regulatory Feedback Networks xRFN beneficial?

slide-21
SLIDE 21

Does the brain perform optimization during recognition? O X O O O O O O O O O O O O O O O O O O O O O E E E E E E E E E E E E E E E E E E E F vs.

Rosenholtz 2001 This occurs in all modalities, including audition, vision, tactile, and (like seen in the jigsaw puzzle)

suggests optimization during recognition is ubiquitous

Rinberg etal 2006

How long does it take to Find the single pattern?

even in olfaction with its poor spatial resolution If brain uses optimization: should be faster with unique patterns (like jigsaw) If brain uses Feedforwad: Y=WX fixed propagation, fixed amount of time (taking what is commonly considered a spatial attention phenomena and attributing it to a recognition phenomena) Brain takes longer in right box suggesting a signal-to-noise phenomena

slide-22
SLIDE 22

Does It Scale To Large AI?

Does Illuminated AI Consume More Resources Than Feedforward AI?

slide-23
SLIDE 23

Tests on Random Data Increasing in Size

Nodes Features Matrix size 10 100 1,000 100 1,000 100,000 500 1,000 500,000 1,000 10,000 10,000,000 2,000 10,000 20,000,000 6,000 12,000 72,000,000 8,000 15,000 120,000,000 9,000 20,000 180,000,000

1,000 2,000 3,000 4,000 5,000 6,000 7,000 Computing Time (s)

Computational Costs:

During Learning

SVM Learning (W) Fastest Feedforward Learning (W) Illuminated Learning (M)

Out of Memory!

5 10 15 20 Matrix Size Million

Out of Memory!

… 120 !

SVM F

Can Learn > 100x Faster Without Balancing Data

Out of Memory!

slide-24
SLIDE 24

2 4 6 8 10 Computational Cost per Test (s)

Computational Costs:

During Recognition

Best Alternate AI (KNN)

Without Optimization

Illuminated AI (M) Matrix Size 50 100 Millions 20 120 Feedforward AI (W)

Nodes Features Matrix size 10 100 1,000 100 1,000 100,000 500 1,000 500,000 1,000 10,000 10,000,000 2,000 10,000 20,000,000 6,000 12,000 72,000,000 8,000 15,000 120,000,000 9,000 20,000 180,000,000 SVM F Out of Memory!

slide-25
SLIDE 25

Accelerated on GPU’s

25

Torch/Lua

slide-26
SLIDE 26

Also Useful for Simpler AI Such As: Logistic Regression & Random Forests

26

slide-27
SLIDE 27

The Current Standard For Explainability Is Decision Trees Based On Logistic Regression

27

30% Loss In Accuracy Occurs When Explaining Using Decision Trees

0% Loss With Adaptive Insight Using Illuminated AI

(In FinTech, Medicine, Government, …)

slide-28
SLIDE 28

Histogram of Factors that Hinder vs. Help the Case

Case #1 Score 0.75 Not Approved Case #2 Score 0.81 Borderline Approved Case #3 Score 0.98 Strongly Approved

Understanding Decisions at a Glance

Number of Factors

  • 0.25 -0.2 -0.15 -0.1 -0.05 0

0.05 0.1 0.15 2 4 6 8 10 12

Helps Hinders

Most Hindering Factor: E With value 66.2 26.8 below expected

2 4 6 8 10 12 14

Helps Hinders

  • 0.5

0.5 1 1.5

Most Helping Factor: G With value -8.8 5.8 above expected

0.6

  • 0.2 -0.1

0.1 0.2 0.3 0.4 0.5 5 10 15 20 25

Helps Hinders

Most Helping Factor : M with score 87.0 12.2 above expected

28

slide-29
SLIDE 29

Comparison:

Structure (during recognition)

Feedforward Illuminated

“Feedforward”

Outputs Y Inputs X

Y X

W

Feedforward-Feedback

Outputs Y Inputs X

Y X

M

M

Explainable?

Yes No

Easy to Learn & Update?

Yes No

Optimization

During Learning During Recognition

1 2 3 4 5 6 7 8 9

29

slide-30
SLIDE 30

Collective Benefits Enabling Wider AI Adoption

Company

Reduce: development costs, time Increase: trust, adoption, and flexibility

Users

Better understanding and trust of AI’s decision process

Developers

Less guessing, easier debugging and updating

Regulators

Understanding of AI’s goals, compromises, and decision process

30

slide-31
SLIDE 31

Offerings

  • Convert & Explain Any Feedforward AI
  • Boost: Internal Development & Quality Assurance
  • Assist Your Regulators: FDA, AMA, DMV, FTC …

1) Illuminate Your AI

  • Faster Learning - 100x
  • Less Data Cleaning
  • User Personalization
  • Easier Update

* for Certain Feedforward AI

2) Train or Update Your AI Fast Without Rehearsal*

slide-32
SLIDE 32

achler@OptimizingMind.com

How Will You Use Illuminated AI?

32

Tsvi Achler 650.486.2303