Dr Juyang Weeng Dr Shaoyun Chen Michigan Sate University

advertisement

Dr Juyang Weeng

Dr Shaoyun Chen

Michigan Sate University

PROS

Efficient for predictable cases

Easier to understand

Computationally inexpensive

CONS

Non generic

Not able to deal with every possible case

Potentially huge number of exhaustive cases.

[1]

Each leaf node represents sample(X,Y)

Each node represents a set of data points with increased similarity

One of the central ideas in Shoslif’s approach

Given X find f (X) at the corresponding leaf node after traversal.

Building a Regression Partition Tree

Take the sample space S.

Divide the space into b cells. Each a child of the root.

The analysis performs automatic derivation of features(discussed later).

Continue to do this until the leaf nodes have a single data point or many data points with virtually the same Y.

8

6

9

2

7

1

3

4 5

2

6

5

1

4

3

9

7

8

Input X’

Output Y control signal

Recursively analyze the centre of each node

If it is close to the input then proceed in that direction till you reach the leaf node .

Use the corresponding Control signal

Use top k paths to find the top k nearest centers.

Feature Selection :Select features from a set of human defined features.

Feature Extraction: extrapolates selected features from images

Feature Derivation : derives features from high dimensional vector inputs

Using Principal Component Analysis recursively partitions the space S into a subspace S’ where the training samples lie.

Computes the principal component vectors .

◦ V

1

,V

2

,V

3

,V

4

…..V

N

MEF : Most Expressive Features

They explain the variation in the sample set

The hyper plane that has V samples forms a partition.

1 as a normal an that passes through the centroid of the

The samples on one side fall onto on side of the tree and vice versa.

PCA LDA

[1]

We can do better with class information.

MDF :Most discriminating feature

Similar to PCA

This method is cuts more along the class boundaries.

Differences

MEF: samples spread out widely, and the samples of different classes tend to mix together.

MDF: samples are clustered more tightly, and the samples from different classes are farther apart.

Using a model similar to Markov chain model

S t

State at time t

At time t, the system is at state S t observes image X t

. and

Control vector Y t+1

S t+1

.

and enters the next state

(S t+1

, Y t+1

) = f (S t

, X t

)

[1]

A special state A (ambiguous) indicates that local visual attention is needed.

Eg. trainer defined this state for a segment right before a turn.

If the image area that revealed the visual difference between different turn types was mainly in a small part of the scene.

A directs the system to look at such landmarks through a prespecified image sub window so that the system issues the correct steering action before it is too late.

Batch learning : All the training data are available at the time the system learns.

Incremental learning : Training samples are available only one at a time.

Discard once you have used them

Memory requires to store the image only once.

Similar images discarded

Step 1 •Query the current RPT.

Step 2

•If the difference between the current RPT’s output and the desired control signal >prespecified tolerance.

•Go to Step 3 Else Goto Step 1

Step 3

•Shoslif learns the current sample to update the RPT.

Compared Shoslif with feed forward neural networks and radial basis function networks for approximating stateless appearancebased navigation systems.

Shoslif did significantly better than both methods.

Extension to face detection, speech recognition and vision-based robot arm action learning.

Shoslif performs better in benign scenes.

The state based method allows more flexibility

However still need to specify that many states for different environment types.

1. Dr.Juyang Weng & Dr. Shaouyun Chen

“Visual Learning with Navigation as an

Example” .Published in IEE

September/October 2000.

Questions

Download