Starting from:
$35

$29

Homework 10 Solution


    1. Model Matching (10 pts): This problem examines the detection of faces by model matching. A face is modeled as a collection of circles and circular arcs, as so:




















The basic idea is that face detection begins by finding circles and circular arcs in the image, followed by matching against a stored model. Be sure to give short justifications for each answer below.

a.    How should the model be represented?

b.  How should images be processed to detect the features?

c.    How should matching be performed?

d.  Is your answer to c. invariant to translation, rotation, and scale?

e.    Can it handle partial occlusion? If so, how? If not, suggest an extension to your scheme that can.

2.  Interpretation Tree (5 pts): We have a choice of matching detected image elements (edges)

to the model or model elements to the object. Let E be the set of detected image edges and M
image, we

|  |

|  |

|  |

|  |

the set of model edges. In the first case, matching image edge to model edges, we generate a
tree of depth

and breadth


with tree size

|  |. In the case of matching the model to

generate a tree of size


. We expect many more image elements than model
elements – there may be many
candidate image edges in a cluttered scene vs. a small number


|  |




of model edges.

    a. Which approach is preferable, matching image edges to model or model to image edges? You might consider the case where there are 12 image edges and 5 model edges, for example.
b.    One advantage to using the interpretation tree approach is that it is possible to match an unknown object in the image to a model even if the object is partially occluded. We do this by allowing an object element to match a “null element” in the model. Does this change your answer to part a.? How and why? Or why not?


| |2
=

2
is the|
−  |

= � # of pixels where



3.  Binary Image Matching (2 pts):
Let 1
and 2 be binary images. Show that








1
2
2

1
2


Where




sum of all (pixels squared) in I.



and
    ⃗
. Assuming that both classes are equally likely, show that the Nearest Mean

    ⃗
4.  Classification (5 pts): Suppose we have a 2-class classification problem with class means

B.


    ⃗

    ⃗






classifier
decision boundary is the hyper-plane perpendicular to, and midway along, the line segment
connecting

to

. You do not need to assume any particular distribution for classes A and

More products