You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: source/py_tutorials/py_feature2d/py_sift_intro/py_sift_intro.rst
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -17,11 +17,11 @@ Theory
17
17
18
18
In last couple of chapters, we saw some corner detectors like Harris etc. They are rotation-invariant, which means, even if the image is rotated, we can find the same corners. It is obvious because corners remain corners in rotated image also. But what about scaling? A corner may not be a corner if the image is scaled. For example, check a simple image below. A corner in a small image within a small window is flat when it is zoomed in the same window. So Harris corner is not scale invariant.
19
19
20
-
.. image:: images/sift_scale_invariant.png
20
+
.. image:: images/sift_scale_invariant.jpg
21
21
:alt:Scale-Invariance
22
22
:align:center
23
23
24
-
So, in 2004, D.Lowe, University of British Columbia, came up with a new algorithm, Scale Invariant Feature Transform (SIFT) in his paper, "Distinctive Image Features from Scale-Invariant Keypoints", which extract keypoints and compute its descriptors. ( This paper is easy to understand and considered to be best material available on SIFT. So this explanation is just a short summary of this paper).
24
+
So, in 2004, **D.Lowe**, University of British Columbia, came up with a new algorithm, Scale Invariant Feature Transform (SIFT) in his paper, **Distinctive Image Features from Scale-Invariant Keypoints**, which extract keypoints and compute its descriptors. *(This paper is easy to understand and considered to be best material available on SIFT. So this explanation is just a short summary of this paper)*.
25
25
26
26
There are mainly four steps involved in SIFT algorithm. We will see them one-by-one.
27
27
@@ -33,13 +33,13 @@ From the image above, it is obvious that we can't use the same window to detect
33
33
But this LoG is a little costly, so SIFT algorithm uses Difference of Gaussians which is an approximation of LoG. Difference of Gaussian is obtained as the difference of Gaussian blurring of an image with two different :math:`\sigma`, let it be :math:`\sigma` and :math:`k\sigma`. This process is done for different octaves of the image in Gaussian Pyramid. It is represented in below image:
34
34
35
35
36
-
.. image:: images/sift_dog.png
36
+
.. image:: images/sift_dog.jpg
37
37
:alt:Difference of Gaussian
38
38
:align:center
39
39
40
40
Once this DoG are found, images are searched for local extrema over scale and space. For eg, one pixel in an image is compared with its 8 neighbours as well as 9 pixels in next scale and 9 pixels in previous scales. If it is a local extrema, it is a potential keypoint. It basically means that keypoint is best represented in that scale. It is shown in below image:
0 commit comments