Description
2 Overview
In this assignment, you will implement a stereo reconstruction algorithm given two view
images.
(a) Left image (b) Right image
Figure 1: In this assignment, you will implement a stereo reconstruction algorithm
given two images.
You can download the skeletal code and data (left.bmp and right.bmp) from here:
https://www-users.cs.umn.edu/~hspark/csci5561/HW5.zip
You will fill main_stereo.m that takes input images and intrinsic parameters K, and
produces a stereo disparity map.
2
CSCI 5561: Assignment #5
Stereo Reconstruction
3 SIFT Feature Matching
(a) Matching from I1 to I2 (b) Matching from I2 to I1
(c) Matching from I1 to I2 after ratio test (d) Matching from I2 to I1 after ratio test
(e) Bidirectional matching between I1 and I2
Figure 2: You will match points between I1 and I2 using SIFT features.
You will use VLFeat SIFT to extract keypoints and match between two views using
k-nearest neighbor search. The matches will be filtered using the ratio test and bidirectional consistency check.
function [x1, x2] = FindMatch(I1, I2)
Input: two input gray-scale images with uint8 format.
Output: x1 and x2 are n × 2 matrices that specify the correspondence.
Description: Each row of x1 and x2 contains the (x, y) coordinate of the point correspondence in I1 ad I2, respectively, i.e., x1(i,:) ↔ x2(i,:). This matching function
is similar to HW#2 except that bidirectional consistency check is mandatory.
(Note) Except for SIFT extraction, you are not allowed to use VLFeat functions.
3
CSCI 5561: Assignment #5
Stereo Reconstruction
4 Fundamental Matrix Computation
Figure 3: Given matches, you will compute a fundamental matrix to draw epipolar
lines.
function [F] = ComputeF(x1, x2)
Input: x1 and x2 are n × 2 matrices that specify the correspondence.
Output: F ∈ R3×3
is the fundamental matrix.
Description: F is robustly computed by the 8-point algorithm within RANSAC. Note
that the rank of the fundamental matrix needs to be 2 (SVD clean-up should be applied.). You can verify the validity of fundamental matrix by visualizing epipolar line
as shown in Figure 3.
(Note) Given the fundamental matrix, you can run the provided function:
[R1 C1 R2 C2 R3 C3 R4 C4] = ComputeCameraPose(F, K)
This function computes the four sets of camera poses given the fundamental matrix
where R1 C1 · · · R4 C4 are rotation and camera center (represented in the world coordinate system) and K is the intrinsic parameter. These four configurations can be
visualized in 3D as shown in Figure 4.
-0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 1.2
0
0.8
0.6
0.4
0.2
0
0.5
0.2
0
0
-0.2
-0.5
-0.4
-1
-0.6
-1.5 0.5
0
-0.5
0
0
0.2
-0.2
0.4
0 0.2 0.4 0.6 0.8 1 -1 -0.5 0 0.5
0.2
0
-0.2
-0.4
-1.5
-0.6
0
Figure 4: Four configurations of camera pose from a fundamental matrix.
4
CSCI 5561: Assignment #5
Stereo Reconstruction
5 Triangulation
Given camera pose and correspondences, you will triangulate to reconstruct 3D points.
function [X] = Triangulation(P1, P2, x1, x2)
Input: P1 and P2 are two camera projection matrices (R3×4
). x1 and x2 are n × 2
matrices that specify the correspondence.
Output: X is n × 3 where each row specifies the 3D reconstructed point.
Description: Use the triangulation method by linear solve, i.e.,
u
1
×
P1
v
1
×
P2
X
1
= 0
(Note) Use plot3 to visualize them as shown in Figure 5.
0 50 100
0
-20
-50
-40
-60
-80
0
0
-100
20
-50
40
60
0
80
50
40
20
40 60 80
0
-20 0 20
-20
-80 -60 -40 0
-40
-20
0
20
0-80 -60 -40 -20 0 20 40 60
(a) nValid = 10 (b) nValid = 488
(c) nValid = 0 (d) nValid = 0
Figure 5: You can visualize four camera pose configurations with point cloud.
5
CSCI 5561: Assignment #5
Stereo Reconstruction
6 Pose Disambiguation
Given four configurations of relative camera pose and reconstructed points, you will
find the best camera pose by verifying through 3D point triangulation.
function [R,C,X] = DisambiguatePose(R1,C1,X1,R2,C2,X2,R3,C3,X3,R4,C4,X4)
Input: R1, C1, X1 · · · R4, C4, X4 are four sets of camera rotation, center, and 3D reconstructed points.
Output: R, C, X are the best camera rotation, center, and 3D reconstructed points.
Description: The 3D point must lie in front of the both cameras, which can be tested
by:
r
T
3
(X − C) > 0 (1)
where r3 is the 3rd row of the rotation matrix. In Figure 5, nValid means the number
of points that are in front of both cameras. (b) configuration produces the maximum
number of valid points, and therefore the best configuration is (b).
6
CSCI 5561: Assignment #5
Stereo Reconstruction
7 Stereo
(a) Rectified image 1 (b) Rectified image 2
Figure 6: Stereo rectification.
Given the disambiguated camera pose, you will implement dense stereo matching between two views based on dense SIFT of VLFeat.
function [H1, H2] = ComputeRectification(K, R, C)
Input: The relative camera pose (R and C) and intrinsic parameter K.
Output: H1 and H2 are homographies that rectify the left and right images such that
the epipoles are at infinity.
Description: Given the disambiguated camera pose, you can find the rectification
rotation matrix, Rrect such that the x-axis of the images aligns with the baseline. Find
the rectification homography H = KRrectRTK−1 where R is the rotation matrix of
the camera. The rectified images are shown in Figure 6. This rectification sends the
epipoles to infinity where the epipolar line becomes horizontal.
(Note) You can use the provided image warping function im_w = WarpImage(im, H)
to check your result.
7
CSCI 5561: Assignment #5
Stereo Reconstruction
Figure 7: Visualization of stereo match.
function [disparity] = DenseMatch(im1, im2)
Input: two gray-scale rectified images with uint8 format.
Output: disparity map disparity ∈ RH×W where H and W are the image height and
width.
Description: Compute the dense matches across all pixels. Given a pixel, u in the
left image, sweep along its epipolar line, lu, and find the disparity, d, that produces the
best match, i.e.,
d = arg min
i
kd
1
u − d
2
u+(i,0)k
2 ∀i = 0, 1, · · · , N
where d
1
u
is the dense SIFT descriptor at u on the left image and d
2
u+(i,0) is the SIFT
descriptor at u+ (i, 0) (i pixel displaced along the x-axis) on the right image. Visualize
the disparity map as shown in Figure 7.
8