A Computational Model of Accelerated Future Learning through Feature Recognition

20
A COMPUTATIONAL MODEL OF ACCELERATED FUTURE LEARNING THROUGH FEATURE RECOGNITION Nan Li Computer Science Department Carnegie Mellon University Building an intelligent agent that simulates human-level learning using machine learning techniques

description

Building an intelligent agent that simulates human-level learning using machine learning techniques. A Computational Model of Accelerated Future Learning through Feature Recognition. Nan Li Computer Science Department Carnegie Mellon University. Accelerated Future Learning. - PowerPoint PPT Presentation

Transcript of A Computational Model of Accelerated Future Learning through Feature Recognition

Page 1: A Computational Model of Accelerated Future Learning through Feature Recognition

A COMPUTATIONAL MODEL OF ACCELERATED FUTURE LEARNING THROUGH FEATURE RECOGNITIONNan LiComputer Science DepartmentCarnegie Mellon University

Building an intelligent agent that simulates human-level learning using machine learning techniques

Page 2: A Computational Model of Accelerated Future Learning through Feature Recognition

ACCELERATED FUTURE LEARNING Accelerated Future Learning

Learning more effectively because of prior learning

Has been observed a lot How?

Expert vs Novice Expert Deep functional feature (e.g. -3x -3) Novice Shallow perceptual feature (e.g. -3x 3)

Page 3: A Computational Model of Accelerated Future Learning through Feature Recognition

A COMPUTATIONAL MODEL Model Accelerated Future Learning Use Machine Learning Techniques Acquire Deep Feature Integrated into a Machine-Learning Agent

Page 4: A Computational Model of Accelerated Future Learning through Feature Recognition

AN EXAMPLE IN ALGEBRA

Page 5: A Computational Model of Accelerated Future Learning through Feature Recognition

FEATURE RECOGNITION ASPCFG INDUCTION Under lying structure in the problem

Grammar Feature Intermediate symbol in a grammar

rule Feature learning task Grammar induction Error Incorrect parsing

Page 6: A Computational Model of Accelerated Future Learning through Feature Recognition

PROBLEM STATEMENT Input is a set of feature recognition records

consisting of An original problem (e.g. -3x) The feature to be recognized (e.g. -3 in -3x)

Output A PCFG An intermediate symbol in a grammar rule

Page 7: A Computational Model of Accelerated Future Learning through Feature Recognition

ACCELERATED FUTURE LEARNING THROUGH FEATURE RECOGNITION Extended a PCFG Learning Algorithm (Li et

al., 2009) Feature Learning Stronger Prior Knowledge:

Transfer Learning Using Prior Knowledge Better Learning Strategy:

Effective Learning Using Bracketing Constraint

Page 8: A Computational Model of Accelerated Future Learning through Feature Recognition

A TWO-STEP ALGORITHM• Greedy Structure

Hypothesizer: Hypothesizes the

schema structure

• Viterbi Training Phase: Refines schema

probabilities Removes redundant

schemas

Generalizes Inside-Outside Algorithm (Lary & Young, 1990)

Page 9: A Computational Model of Accelerated Future Learning through Feature Recognition

GREEDY STRUCTURE HYPOTHESIZER Structure learning Bottom-up Prefer recursive to non-recursive

Page 10: A Computational Model of Accelerated Future Learning through Feature Recognition

EM PHASE Step One:

Plan parse tree computation

Most probable parse tree

Step Two: Selection

probabilities update s: ai p, aj ak

Page 11: A Computational Model of Accelerated Future Learning through Feature Recognition

FEATURE LEARNING Build Most Probable

Parse Trees For all observation

sequences Select an

Intermediate Symbol that Matches the most

training records as the target feature

Page 12: A Computational Model of Accelerated Future Learning through Feature Recognition

TRANSFER LEARNING USING PRIOR KNOWLEDGE GSH Phase:

Build parse trees based on previously acquired grammar

Then call the original GSH

Viterbi Training: Add rule frequency

in previous task to the current task

0.66

0.330.50.5

Page 13: A Computational Model of Accelerated Future Learning through Feature Recognition

EFFECTIVE LEARNING USING BRACKETING CONSTRAINT Force to generate a

feature symbol Learn a subgrammar

for feature Learn a grammar for

whole trace Combine two

grammars

Page 14: A Computational Model of Accelerated Future Learning through Feature Recognition

EXPERIMENT DESIGN IN ALGEBRA

Page 15: A Computational Model of Accelerated Future Learning through Feature Recognition

EXPERIMENT RESULT IN ALGEBRA

Fig.2. Curriculum one Fig.3. Curriculum two Fig.4. Curriculum three

Both stronger prior knowledge and a better learning strategy can yield accelerated future learning

Strong prior knowledge produces faster learning outcomes L00 generated human-like errors

Page 16: A Computational Model of Accelerated Future Learning through Feature Recognition

LEARNING SPEED INSYNTHETIC DOMAINS

Both stronger prior knowledge and a better learning strategy yield faster learning

Strong prior knowledge produces faster learning outcomes with small amount of training data, but not with large amount of data

Learning with subtask transfer shows larger difference, 1) training process; 2) low level symbols

Page 17: A Computational Model of Accelerated Future Learning through Feature Recognition

SCORE WITH INCREASING DOMAIN SIZES

The base learner, L00, shows the fastest drop Average time spent per training record

Less than 1 millisecond except for L10 (266 milliseconds) L10: Need to maintain previous knowledge, does not separate trace

into small traces Conciseness: Transfer learning doubled the size of the schema.

Page 18: A Computational Model of Accelerated Future Learning through Feature Recognition

INTEGRATING ACCELERATED FUTURE LEARNING IN SIMSTUDENT

Tutor LuckyNext Problem Quiz Lucky

Prepare Lucky for Quiz Level 3 !

Curriculum Browser

Level 1:[+] One-Step Linear Equation

Level 2:[+] Two-Step Linear Equation

Level 3:[-] Equation with Similar Terms

OverviewIn this unit, you will solve equations with integer or decimal coefficients, as well as equations involving more than one variable.

More…

Lucky

x+5

• A machine-learning agent that• Acquires

production rules from

• Examples and problem solving experience

• Integrate the acquired grammar into production rules Requires weak

operators (non-domain specific knowledge)

Less number of operators

Page 19: A Computational Model of Accelerated Future Learning through Feature Recognition

CONCLUDING REMARKS Presented a computational model of human

learning that yields accelerated future learning.

Showed Both stronger prior knowledge and a better

learning strategy improve learning efficiency. Stronger prior knowledge produced faster

learning outcomes than a better learning strategy.

Some model generated human-like errors, while others did no make any mistake.

Page 20: A Computational Model of Accelerated Future Learning through Feature Recognition