Sale!

CS6476 Problem Set 5: Object Tracking and Pedestrian

$35.00 $21.00

Category: You will Instantly receive a download link for .zip solution file upon Payment || To Order Original Work Click Custom Order?

Description

5/5 - (5 votes)

Detection
Description
In this problem set you are going to implement tracking methods for image sequences and videos. The
main algorithms you will be using are the Kalman and Particle Filters.
Learning Objectives
● Identify which image processing methods work best in order to locate an object in a scene.
● Learn to how object tracking in images works.
● Explore different methods to build a tracking algorithm that relies on measurements and a prior
state.
● Create methods that can track an object when occlusions are present in the scene.
Problem Overview
Methods to be used: You will design and implement a Kalman and a Particle filters from the ground up.
RULES: You may use image processing functions to find color channels, load images, find edges (such as
with Canny), resize images. Don’t forget that those have a variety of parameters and you may need to
experiment with them. There are certain functions that may not be allowed and are specified in the
assignment’s autograder Piazza post.
Refer to this problem set’s autograder post for a list of banned function calls.
Please do not use absolute paths in your submission code. All paths should be relative to the
submission directory. Any submissions with absolute paths are in danger of receiving a penalty!
Obtaining the Starter Files:
Obtain the starter code from canvas under files.
Programming Instructions
Your main programming task is to complete the api described in the file ps5.py. The driver program
experiment.py helps to illustrate the intended use and will output the files needed for the writeup.
Write-up instructions
Create ps5_report.pdf – a PDF file that shows all your output for the problem set, including images
labeled appropriately (by filename, e.g. ps5-1-a-1.png) so it is clear which section they are for and the
small number of written responses necessary to answer some of the questions (as indicated). For a
guide as to how to showcase your results, please refer to the powerpoint template for PS5.
How to submit:
1. To submit your code, in the terminal window run the following command:
python submit.py ps05
2. To submit the report, input images for part 5, and experiment.py, in the terminal window run
the following command:
python submit.py ps05_report
3. Submit your report pdf to gradescope.
YOU MUST PERFORM ALL THREE STEPS. I.e. two commands in the terminal window and one upload to
gradescope. Only your last submission before the deadline will be counted for each of the code and
the report.
The following lines will appear:
GT Login required.
Username :
Password:
Save the jwt?[y,N]
You should see the autograder’s feedback in the terminal window. Additionally, you can look at a history
of all your submissions at https://bonnie.udacity.com/
Grading
The assignment will be graded out of 100 points. The last submission before the time limit will only be
considered. The code portion (autograder) represents 50% of the grade and the report the remaining
50%.
The images included in your report must be generated using experiment.py. This file should be set to be
run as is to verify your results. Your report grade will be affected if we cannot reproduce your output
images.
The report grade breakdown is shown in the question heading. As for the code grade, you will be able to
see it in the console message you receive when submitting.
Assignment Overview
1. The Kalman Filter [20 points]
The Kalman Filter (KF) is a method used to track linear dynamical systems with gaussian noise. Recall
that both the predicted and the corrected states are gaussians, allowing us to keep the mean and
covariance of the state in order to run this filter. Use the kalman filter on the provided images to track a
moving circle.
First we need to define the KF’s components. We will use the notation presented in the lectures for the
N-dimensional case.
State vector:
We represent the filter’s state using the current position and velocities in the x, y plane. We will not use
acceleration in our exercise.
X = [x, y, v , ]
x vy
Dynamics Model transition 4×4 matrix D :
t
This matrix will map the equations below to the state components. Determine how this 4×4 matrix can
be used to represent these equations using the state vector.
x Δt
t = xt−1 + vx y Δt
t = yt−1 + vy
Δt = 1 for simplicity
State and Covariance prediction:
We assume that the state vector prediction is based on a linear transformation using the transition
matrix D . And the covariance is obtained from the squaring operation plus the process noise . We t Σdt
will also assume this noise is independent from each component in the state vector.
Xt X
− = Dt
+
t−1 Σt Σ D Σ
− = Dt
+
t−1 t
T + dt
Before we continue to the Correction state. We will define our sensor’s function as a way to obtain the
object’s position. In our case we will use a template matching code to obtain these components. In
order to make this a little more interesting, we will add some noise to these measurements.
Sensor measurement 2×4 matrix M :
t
This matrix maps the available measurements to the state vector defined above. Because we can only
obtain two values, x and y , it contains two rows and four columns.
Kalman Gain K :
t
Using the measurement matrix, and the predicted covariance we can now compute the Kalman Gain.
Here you will also define a measurement noise represented by Σm .
t
K M (M Σ M )
t = Σt

t
T
t t

t
T + Σmt
−1
State and Covariance correction:
Now that we have all the components we need, you will proceed to update the state and the covariance
arrays. Notice that in order to update the state you will use the residual represented by the difference
between the measurements obtained from the sensor (template matching function) Y and the t
predicted positions M X .
t t

X (Y X )
t
+ = Xt
− + Kt t − Mt t
− Y are the measurements at time t
t
Σt
I M ) Σ
+ = ( − Kt t t

I is the identity matrix
In the next iteration the updated state and covariance will be used in the prediction stage.
This may seem like a lot but you will find coding this KF is not complicated. Complete the
KalmanFilter class with the process described above. We will split the tasks the following way:
a. Initialize the class defining the class arrays you will be using. Place these components in the
__init__(self, init_x, init_y) function. Here’s a reminder of what should be included in this part:
i. State vector ( X ) with the initial x and y values.
ii. Covariance 4×4 array ( Σ ) initialized with a diagonal matrix with some value.
iii. 4×4 state transition matrix Dt
iv. 2×4 measurement matrix Mt
v. 4×4 process noise matrix Σdt
vi. 2×2 measurement noise matrix Σm t
Code: __init__(self, init_x, init_y)
Complete the prediction state in predict(self). Here you will replace the class variables for the
state and covariance arrays with the prediction process.
Code: predict(self)
Finally, we need to correct the state and the covariances using the Kalman gain and the
measurements obtained from our sensor.
Code: correct(self)
b. The position estimation is performed by the function process(self, measurement_x,
measurement_y) which uses the functions defined above. Now that you have finished creating
the Kalman Filter, estimate the position of a bouncing circle. Use the images in the directory
called ‘circle’ to find the circle’s position at each frame. Include the frame numbers described
below in your report.
Report: Frames to record: 10, 30, 59, 99
– Input: circle/0000.jpg to circle/0099.jpg.
– Output: ps5-1-b-1.png, ps5-1-b-2.png, ps5-1-b-3.png, ps5-1-b-4.png
c. Now let’s try with a more complicated scene using the same filter. We will find and track
pedestrians in a video. This means we need to update our sensor with a function that can locate
people in an image. Fortunately, there is a function in OpenCV that can help us with this task
using Histogram of Gradients (HoG) descriptors. This object contains a method called
detectMultiScale that returns the bounding boxes for each person in the scene along with the
detection weight associated to it.
Use the sequence of images in the ‘walking’ directory to detect and follow the man crossing the
street using the output from the HoG detector as the measurements and the Kalman Filter.
Report: Frames to record: 10, 33, 84, 159
– Input: walking/001.jpg to walking/159.jpg.
– Output: ps5-1-c-1.png, ps5-1-c-2.png, ps5-1-c-3.png, ps5-1-c-4.png
2. The Particle Filter [25 points]
Recall that we need a variety of elements:
(1) a model – this is the “thing” that is actually being tracked. Maybe it’s a patch, a contour, or some
other description of an entity;
(2) a representation of state x that describes the state of the model at time ;
t
t
(3) a dynamics model p(x | x ) that describes the distribution of the state at time given the t t−1
t
state at t − 1 ;
(4) a measurement z that somehow captures the data of the current image; and finally,
t
(5) a sensor model p(z | x ) that gives the likelihood of a measurement given the state. For
t t
Bayesian-based tracking, we assume that at time t − 1 we have a belief about the state
represented as a density p(x ) , and that given some measurements at time we update t−1
zt
t
our Belief by computing the posterior density:
Bel(xt) ∝ p(zt|xt)p(xt|ut,xt-1)Bel(xt-1)
In class we discussed both the theory and the implementation specifics behind particle filtering. The
algorithm sketch is provided in the slides and it describes a single update of the particle filter. What is
left up to the coder are several details including what is the model, what is the state, the number of
particles, the dynamics/control function p(x | x , u ) , the measurement function , and the t t−1 t p(z | x )
t t
initial distribution of particles.
For this question, you will to need track an image patch template taken from the first frame of the
video. For this assignment the model is simply going to be the image patch, and the state will be only the
2D center location of the patch. Thus each particle will be a (u, v) pixel location (where u and v are the
row and column number respectively) representing a proposed location for the center of the template
window.
We will be using a basic function to estimate the dissimilarity between the image patch and a window of
equal size in the current image, the Mean Squared Error:
The funny indexing is just because (u, v) are indexed from 1 to M (or N) but the state is the
location of the center. Notice the above equation can be simplified by using array operations.
Remember you are calculating the Mean of the Squared Error between two images, a template and a
patch of the image.
These images are color and you will convert them to grayscale can do the tracking in either the full color
image, or in grayscale, or even in the green channel. If you use color, there would be an outer
summation over the three color channels. If you want to use grayscale, you can use cvtColor in OpenCV
or just use a weighted sum of R,G,B to get a luma value (try weights of 0.3, 0.58, and 0.12).
MSE, of course, only indicates how dissimilar the image patch is, whereas we need a similarity measure
so that more similar patches are more likely. Thus, we will use a squared exponential equation
(Gaussian) to convert this into a usable measurement function:
To start out, you might use a value of σMSE = 10 (for grayscale in 0-255 range) but you might need to
change this value.
For the dynamics we’re going to use normally distributed noise since the object movements can be
unpredictable and often current velocity isn’t indicative of future motion; so our dynamics model is
merely
which is just a fancy way to say you add a little bit of Gaussian noise to both u and v independently.
The number of particles and initial distribution are up to you to figure out. You may need to tune your
parameters for these to get better tracking/performance.
In order to visualize the tracker’s behavior you will need to overlay each successive frame with the
following elements:
– Every particle’s (x,y) location in the distribution should be plotted by drawing a colored dot
point on the image. Remember that this should be the center of the window, not the corner.
– Draw the rectangle of the tracking window associated with the Bayesian estimate for the
current location which is simply the weighted mean of the (x,y) of the particles.
– Finally we need to get some sense of the standard deviation or spread of the distribution. First,
find the distance of every particle to the weighted mean. Next, take the weighted sum of these
distances and plot a circle centered at the weighted mean with this radius.
You will have to produce these visuals for select frames but you will not have to submit the entire video.
For reading the frames OpenCV users should look at the class VideoCapture. For sampling based on a
distribution, see the functions numpy.random.multinomial or numpy.random.choice.
a. Implement the particle filter. You can find the bounding box for the initial position in
template_rect – this is your template. Tweak the parameters including window size until you can
get the tracker to follow this object. Run the tracker and save the video frames described below
with the visualizations overlaid.
Classes, methods: Define a class ParticleFilter, with the following methods:
– __init__(frame, template, **kwargs): (constructor) Initialize particle filter object.
– get_error_metric(template, frame_cutout): Calculates the error metric used (i.e. MSE).
– resample_particles(): Returns a list of particles resampled based on their weights.
– process(frame): Process a frame (image) of video and update filter state.
* Hint: You should call resample_particles() and get_error_metrics() here.
– render(frame_out): Visualize current particle filter state.
A helper function, run_particle_filter(), has been provided to run your particle filter on
a video and collect the output. You can also display every video frame to verify that your particle
filter is working properly.
Note: The helper function run_particle_filter() can save the template image patch and
outputs generated by your render() method for desired frames. See template code for details.
Report: Frames to record: 10, 30, 59, 99
– Input: circle/0000.jpg to circle/0099.jpg.
– Output: ps5-2-a-1.png, ps5-2-a-2.png, ps5-2-a-3.png, ps5-2-a-4.png
b. Now use the particle filter with a noisy video:
Report: Frames to record: 10, 33, 84, 99
– Input: pres_debate_noisy/000.jpg to pres_debate_noisy/099.jpg.
– Output: ps5-2-b-1.png, ps5-2-b-2.png, ps5-2-b-3.png, ps5-2-b-4.png
3. Changes in Appearance [10 points]
In the last section you were working with a static template assuming that the target will not change in
shape. This can be addressed by updating the model’s appearance over time.
Modify your existing tracker to include a step which uses the history to update the tracking window
model. We can accomplish this using what’s called an Infinite Impulse Response (IIR) filter. The concept
is simple: we first find the best tracking window for the current particle distribution as displayed in the
visualizations. Then we just update the current window model to be a weighted sum of the last model
and the current best estimate.
where Best(t) is the patch of the best estimate or mean estimate. It’s easy to see that by recursively
updating this sum, the window implements an exponentially decaying weighted sum of (all) the past
windows. For this assignment, you may assume t to be the frame number, in lieu of a time measure.
a. Implement the appearance model update feature. Run the tracker on the presidential debate
images and adjust parameters until you can track the person’s hand (the one not holding the
microphone). Your tracker’s bounding box should only contain the hand at all times. Run the
tracker and save the video frames described below with the visualizations overlaid.
Classes, methods: Derive a class AppearanceModelPF from ParticleFilter, retaining the
same interface – i.e., with methods process() and render(), as defined above. The constructor
and process() method should transparently handle model updates. You can supply additional
keyword arguments to the constructor, if needed.
Report: Frames to record: 22, 50, 160
– Input: pres_debate/000.jpg to pres_debate/165.jpg.
– Output: ps5-3-a-1.png, ps5-3-a-2.png, ps5-3-a-3.png
4. Particle Filters and Occlusions [20 points]
For this part we will work with a much more difficult video to perform tracking with, pedestrians.mp4.
We’d like to be able to track the blond-haired woman (the one with the white jacket) as she crosses the
road. If you try applying your adaptive tracker to this video, you will probably find that you will have
difficulty dealing simultaneously with occlusion and the perspective shift as the woman walks away from
the camera. Thus, we need some way of relying less on updating our appearance model from previous
frames and more on a more sophisticated model of the dynamics of the figure we want to track.
Expand your appearance model to include window size as another parameter. This will change the
representation of your particles. You are highly recommended to use cv2.resize for your implementation
to resize the template to track.
a. Run the tracker and save the video frames described below with the visualizations overlaid. You
will receive full credit if you can reliably track (illustrate this with the rectangle outline) all the
way to the end of the street and deal gracefully with the occlusions (reasonable tracking at
frame 300). The tracking box must track the person’s position and size.
Classes, methods: Derive a class MDParticleFilter from AppearanceModelPF, retaining the
same interface – i.e., with methods process() and render(), as defined above. The constructor
and process() method should transparently handle model updates in appearance and dynamics.
You can supply additional keyword arguments to the constructor and other methods, if needed.
Report: Frames to record 40, 100, 240, 300.
– Input: pedestrians/000.jpg to pedestrians/319.jpg.
– Output: ps5-4-a-1.png, ps5-4-a-2.png, ps5-4-a-3.png, and ps5-4-a-4.png
– Text Answer: Describe what you did. How did you modify the Particle Filter class to continue
tracking after occlusions?
5. Tracking Multiple Targets [20 points]
Now use either a kalman or particle filter to track multiple targets as they move through the given video.
Use the sequence of images in the TUD-Campus directory to detect the people shown below.
Report: Frames to record: 29, 56, 71
– Input: TUD-Campus/01.jpg to TUD-Campus/71.jpg.
– Output: ps5-5-a-1.png, ps5-5-a-2.png, ps5-5-a-3.png
– Text Answer: Describe what you did. How different it was to use a KF vs PF? Which one worked best
and why? Include details about any modifications you had to apply to handle multiple targets.
6. Challenge Problem: Detect Pedestrians from a moving camera [5 points]
Finally, detect pedestrians from the given video of a moving camera. Use the images in the directory
‘follow’ to track the man with the hat holding a white plastic bag.
Report: Frames to record: 60, 160, 186
– Input: follow/001.jpg to follow/186.jpg.
– Output: ps5-6-a-1.png, ps5-6-a-2.png, ps5-6-a-3.png
– Text Answer: Describe what you did. Did this task present any additional challenges compared to the
previous sections? Include details about any modifications you had to apply.