Professional Documents
Culture Documents
Feature Descriptors: Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab
Feature Descriptors: Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab
Feature Descriptors: Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab
Feature Descriptors
normalized region
e.g. color e.g. color
• Fast approximation
– Avoid computing the
eigenvalues
– α: constant “Flat” “Edge”
(0.04 to 0.06) region θ<0
l1
Stanford University Lecture 7 - 7 17-Oct-17
Quick review: Harris Corner Detector
• Translation invariance
• Rotation invariance
• Scale invariance?
f Image 1 f Image 2
scale = 1/2
f Image 1 f Image 2
scale = 1/2
L = s ( Gxx ( x, y, s ) + Gyy ( x, y, s ) )
2
(Laplacian)
DoG = G( x, y, ks ) - G( x, y, s )
(Difference of Gaussians)
where Gaussian
x2 + y 2
- Note: both kernels are invariant
G ( x, y , s ) = 1
2ps
e 2s 2
to scale and rotation
¬ Laplacian ®
Find local maximum of:
– Harris corner detector in
y
space (image coordinates)
– Laplacian in scale ¬ Harris ® x
1 K.Mikolajczyk, C.Schmid. Indexing Based on Scale Invariant Interest Points . ICCV 2001
2 D.Lowe. Distinctive Image Features from Scale-Invariant Keypoints . IJCV 2004
¬ Laplacian ®
Find local maximum of:
– Harris corner detector in
y
space (image coordinates)
– Laplacian in scale ¬ Harris ® x
• SIFT (Lowe)2 scale
Find local maximum of:
– Difference of Gaussians in space
¬ DoG ®
and scale
y
¬ DoG ® x
1 K.Mikolajczyk, C.Schmid. Indexing Based on Scale Invariant Interest Points . ICCV 2001
2 D.Lowe. Distinctive Image Features from Scale-Invariant Keypoints . IJCV 2004
( )
G k 2s * I
G(ks )* I D(s ) º (G(ks ) - G(s ))* I
G(s )* I
( )
D k 2s
D(ks )
D(s )
source
source
Repeatability rate:
# correspondences
# possible correspondences
K.Mikolajczyk, C.Schmid. Indexing Based on Scale Invariant Interest Points . ICCV 2001
Following slides credit: CVPR 2003 Tutorial on Recognition and Matching Based on Local
Invariant Features David Lowe
Stanford University Lecture 7 - 32 17-Oct-17
Advantages of invariant local features
• Locality: features are local, so robust to
occlusion and clutter (no prior segmentation)
• Distinctiveness: individual features can be
matched to a large database of objects
• Quantity: many features can be generated for
even small objects
• Efficiency: close to real-time performance
• Extensibility: can easily be extended to wide
range of differing feature types, with each
adding robustness
34
Stanford University Lecture 7 - 17-Oct-17
Becoming rotation invariant
0 2p
• Using precise gradient locations is fragile. We’d like to allow some “slop” in
the image, and still produce a very similar descriptor
• Create array of orientation histograms (a 4x4 array is shown)
• Put the rotated gradients into their local orientation histograms
– A gradients’s contribution is divided among the nearby histograms based on distance. If
it’s halfway between two histogram locations, it gives a half contribution to both.
– Also, scale down gradient contributions for gradients far from the center
• The SIFT authors found that best results were with 8 orientation bins per
histogram.
Vectors indicate
scale, orientation
and location.
• Using precise gradient locations is fragile. We’d like to allow some “slop” in
the image, and still produce a very similar descriptor
• Create array of orientation histograms (a 4x4 array is shown)
• Put the rotated gradients into their local orientation histograms
– A gradients’s contribution is divided among the nearby histograms based on distance. If
it’s halfway between two histogram locations, it gives a half contribution to both.
– Also, scale down gradient contributions for gradients far from the center
• The SIFT authors found that best results were with 8 orientation bins per
histogram, and a 4x4 histogram array.
David G. Lowe, "Distinctive image features from scale-invariant keypoints," International Journal of
Computer Vision, 60, 2 (2004), pp. 91-110
David G. Lowe, "Distinctive image features from scale-invariant keypoints," International Journal of
Computer Vision, 60, 2 (2004), pp. 91-110
http://www.cs.ubc.ca/~mbrown/autostitch/autostitch.html
Stanford University Lecture 7 - 52 17-Oct-17
Wide baseline stereo
17-Oct-17
Stanford University 58
Lecture 7-
Histogram of
Oriented Gradients
• Dividing the image window into small
spatial regions (cells)
• Cells can be either rectangle or radial.
• Each cell accumulating a weighted local
1-D histogram of gradient directions
over the pixels of the cell.
17-Oct-17
Stanford University Lecture 7 - 59
Histogram of Oriented Gradients
• Six variables
– each point give two equations
– at least three points
• Least squares