Discovering Objects in Videos

Transcription

Discovering Objects in Videos
Discovering Objects in Videos
Karim Sayed Ahmed
[email protected]
Machine learning project - Final Report
1
Problem
The goal of this project is to identify and localize objects in videos. Given a video V ,
consisting of a set of N frames f1 , .., fN where each frame contains one or more unknown
object, the objective is to identify and localize primary objects in each i-th frame fi in V
that exhibit coherence in both appearance and motion. Figure 1 shows a brief illustration of
discovering objects in videos problem. There are two sub-problems that should be solved;
the first sub-problem is localization of primary objects in each frame (still images). The
second sub-problem is finding the correlation between the detected objects in all video’s
frames.
F1
FN-1
F2
Primary Object ?
FN
Same Objects ?
Figure 1: The problem of discovering objects in videos is divided into two sub-problems. The first
sub-problem is localization of primary objects in each frame. The second sub-problem is finding the
correlation between the detected objects in all video’s frames.
2
Related Work
The problem of discovering objects in videos is closely related to video segmentation. The
following are some of the state-of-the-art related work in this area.
• Representing video by a multi-label Markov Random Field model, and
accomplishing segmentation by finding the minimum energy label [7].
• Fragments-based tracking of non-rigid objects using level sets [2].
• Using layered Directed Acyclic Graph (DAG) based framework for detection and
segmentation of the primary object in video [10].
3
Approach Overview
In order to identify and localize coherent objects in videos, our proposed approach is
divided into two main phases, where each phase solves one of the sub-problems illustrated
in Figure 1. The key contribution of this work is utilizing deep learning and transductive
learning to discover video objectness. The following subsections discusses an overview of
the proposed two-phases approach.
1
3.1
Phase I: Object localization per frame
The objective of this phase is to find the primary object in each frame in the video; this
primary object is represented in a bounding box. To accomplish this task, I used an
approach called STL which is proposed in [1] and it relies on deep learning. This approach
is a new self-taught object localization algorithm for still images that leverages on deep
convolutional neural networks trained for whole image recognition to localize objects
in images without additional human supervision, i.e., without using any ground-truth
bounding boxes for training [1]. This approach can generate many bounding boxes
containing object proposals. After extracting the object proposals in each frame, the next
step is to re-rank the detected top four object proposals using a linear SVM. According
to our experiments, we found that this step is essential and it enhances the quality of the
detected video’s objects as it will be shown later. Figure 2 gives an overview of the steps
included in phase I.
Top four object proposals
(using deep learning STL [1])
The best object Proposal
(using SVM)
Figure 2: Overview of phase I.
3.2
Phase II: Classification of the primary objects in video
After identifying and localizing the primary object in each frame (output of phase I ), the
next step is classification of the primary objects in all video’s frames using Transductive
Support Vector Machine (TSVM). This learning problem could be formulated as detection
of similarity between objects; discarding noise objects.
The are two main issues with learning the similarity between frames of the same video
which are: 1) Few labeled samples in training data especially for positive samples. 2) The
testing data are part of the training data.
By using Transductive Support Vector Machine (TSVM) [4], we can avoid these problems.
The advantage of using Transductive Support Vector Machines is that it trains labeled and
unlabeled data, and the unlabeled data is part of the test data. In other words, TSVM
extends SVM by treating partially labeled data in semi-supervised learning by following
the principles of transduction.
In addition to the training set D, the learner is also given a set D? = {x?i |x?i ∈ Rp }ki=1
of test examples to be classified. Transductive support vector machine is defined by the
following primal optimization problem [4] as follows:
Minimize (over w, b, y? ) : 21 kwk2 ,
subject to (for any i = 1, . . . , n and any j = 1, . . . , k):
yi (w · xi − b) ≥ 1,
yj? (w · x?j − b) ≥ 1,
and
yj? ∈ {−1, 1}.
2
4
Dataset and extraction of image features
For testing and evaluation, we used the SegTrack dataset [7, 8]. It consists six videos
with pixel-level segmentation ground-truth for each video. The standard method used
for evaluation in this dataset is the average per-frame pixel error. This dataset is widely
used by other approaches including some of the recent proposed methods [5, 10, 7, 8], so it
is feasible to compare our results with other approaches. For extracting image features, we
used the Histogram of Gradients (HOGs [3]).
Figure 3: Videos used by Segtrack [7] dataset.
5
Learning and Inference
In this section, we will discuss the various methods and techniques used in the learning
and inference process in the two-phases proposed approach.
5.1
Learning and Inference for Phase I
First, and before going into the details of the learning and inference process. We have to
show why we need extra learning step after executing STL. As mentioned earlier, in this
part I applied STL approach [1] on Segtrack dataset. The testing code for STL has been
made available by the authors of paper [1]. However, selecting the top-1 bounding box
by ranked STL, is not a good option, as it will be shown later in the experimental results
section. But for now, we illustrate Figure 4 which shows a visualization of some video
frames applied on Segtrack dataset, which will bounding box (object proposal) with the
highest confidence score as it is most likely will contain the primary object.
Figure 4: Results of applying STL on Segtrack video frames. Red bounding box in each frame is the
selected Top-1 score object proposal. Average error rate = 64%
It is clear from the results in Figure 4 that the top-1 bounding box miss-classifies the
primary object in the scene many times. My experiments that are conducted on Segtrack
dataset show that the average error rate for selecting STL top-1 bounding box = 64%. In
order to get good results, it is crucial for this phase to select the best possible bounding
3
boxes as the overall performance will depend heavily on these results. To tackle this
problem, we investigated different object proposals generated by STL. Figure 5 shows the
Top-4 bounding boxes generated by STL.
(a) Although red bounding box (Top-1) has the (b) Although red bounding box (Top-1) has the
highest STL confidence score, Top-4 and Top-2 are highest STL confidence score, Top-4 and Top-3 are
much better.
much better.
Figure 5: Top four bounding boxes from STL. Red box is the Top-1, other blue boxes are Top-2,
Top-3, and Top-4.
In Figure 5, STL showed overall good semantic results; however, it is obvious that STL
tends to give higher score for the bigger bounding boxes. This raised a serious problem
that will definitely affect the final project results. To overcome this problem, we have to
execute an extra learning process on the top ranked bounding boxes generated by STL as
it will be shown in the following subsections.
In this phase, we used a linear SVM classifier to re-rank the STL object proposals for each
frame. We propose two different methods for applying SVM. The first method is learning
SVM classifier on Segtrack dataset; and the second method is learning SVM classifier on
top-1 STL bounding boxes. In the following parts, we discuss these two methods in details.
5.1.1
Method(1): Learning with linear SVM on Segtrack ground-truth
To refine the ranking score of STL, we applied linear SVM to re-rank the top-four object
proposals. There are two issues to apply this solution. First, SVM is a two-classes
discriminative learning algorithm. Second, the availability of ground truth dataset. The
following is my proposed and implemented solution to these problems.
First, applying SVM to re-rank object proposals for every frame can be seen as detecting
the object proposal which has the highest abjectness value. In other words, it aims at
discriminating objects from non-objects in every frame. So, class (+1) is assigned to ground
truth object bounding box; that is all objects are alike and there is no difference between
different objects. On the other hand, class (-1) is assigned to every thing else (any non
object bounding box). The negative examples are extracted from the background of each
frame.
Second, as mentioned in the previous paragraph, we have to find a ground-truth dataset
to get the perfect tight bounding box that contains only the primary object in the frame.
Fortunately, the Segtrack dataset includes a segmented ground truth images for each frame
in videos. Figure 6 shows the process of extracting perfect objects (ground truth); and
applying SVM on one frame. At the training time, the perfect bounding box is considered
class (+1) and parts from frame background are considered negative examples class (-1).
5.1.2
Method(2): Learning with linear SVM on top-1 STL
This method is an alternative for Method (1). In this method, we also used a linear SVM
to re-rank the top four STL objects proposals; however in we trained SVM on the top-1
STL bounding box, extracted on all frames for all videos. Figure 7 shows the process of
selecting the top-a STL bounding box as ground truth; and then applying a linear SVM on
4
-1
+1
Segtrack ground-truth
Figure 6: Method(1) Training. Perfect objects are extracted from Segtrack dataset (ground truth),
then we train a linear SVM classifier on these objects in all frames through all the videos. At the
training time, the perfect bounding box is a positive example (class (+1)), and background parts are
a negative (class (-1)).
all boxes. At the training time, the top-1 STL bounding box is considered class (+1) and
parts from frame’s background are considered negative examples class (-1).
Top-1 STL
-1
+1
Figure 7: Method(2) Training. Ground-truth objects considered the top-1 STL bounding boxes,
extracted from Segtrack dataset. We train a linear SVM classifier on these top-1 objects in all frames
through all the videos. At the training time, the top-1 STL bounding box is a positive example (class
(+1)), and background parts are a negative (class (-1)).
5.1.3
Testing for Phase I
After learning a linear SVM classifier as illustrated whether in Method(1) or Method(2), the
next step is to test the trained model on the different STL bounding boxes in each frame,
and for each video. Figure 9 shows testing the SVM model (Method(1) or Method(2)) on
the top four STL bounding boxes. The highest ranked bounding box is selected to be the
primary object in the frame.
Test SVM on Top four STL
Select highest score
Figure 8: Testing for Phase I
5
5.2
Learning and Inference for Phase II
After identifying and localizing the primary object in each frame (output of phase I ), the
next step is classification of the primary objects in all video’s frames using Transductive
Support Vector Machine (TSVM). This learning problem is considered as detection of
similarity between objects. The are two main issues with learning the similarity between
frames of the same video which are: 1) Few labeled samples in training data especially
for positive samples. 2) The testing data are part of the training data. The advantage of
using TSVM is that it trains labeled and unlabeled data, and the unlabeled data is part
of the test data. In other words, TSVM extends SVM by treating partially labeled data
in semi-supervised learning. Figure 9 shows the learning and inference process using
TSVM. The first primary object in the first frame in the video is considered the only positive
example (shown in red rectangle, class(+1)); the negative examples are extracted from the
background parts of the video’s frames; and the other primary objects (shown in yellow
rectangles) in the other frames are considered the unlabeled data ans assigned value of
zeros (unlabeled).
-1
+1
0: (unlabeled)
Figure 9: Learning and inference using Transductive Support Vector Machine (TSVM) for phase II.
The first primary object in the first frame in the video is the only positive example (shown in red
rectangle, class(+1)); the negative examples are extracted from the background parts; and the other
primary objects (shown in yellow rectangles) in the other frames are the unlabeled data.
6
Experimental Results
In this section, we discuss the overall performance of our proposed two-phases approach,
including the two proposed methods “Method (1)” and “Method (2)” (with and without
TSVM). In addition to illustrating experimental results for the other possible methods.
Before discussing the details in the next subsections, we first provide a legend for the
names of the different proposed and implemented methods as follows:
• “Method(1) with TSVM“: This method uses a linear SVM classifier to re-rank
the top four STL bounding boxes in phase I. The ground-truth training dataset
is the segmented object included in Segtrack dataset. Then, in phase II, a TSVM
classifier is applied to find similar objects in each video.
• “Method(2) with TSVM“: This method uses a linear SVM classifier to re-rank the
top four STL bounding boxes in phase I. The ground-truth training dataset is the
top-1 STL bounding according to the ranking of STL . Then, in phase II, a TSVM
classifier is applied to find similar objects in each video.
• “Method(1) without TSVM“: This method uses only a linear SVM classifier to
re-rank the top four STL bounding boxes. The ground-truth training dataset is the
segmented object included in Segtrack dataset. There is no phase II.
• “Method(2) without TSVM“: This method uses only a linear SVM classifier to
re-rank the top four STL bounding boxes. The ground-truth training dataset is the
6
top-1 STL bounding according to the ranking of STL . There is no phase II.
• “Method(3)“: This is a method in which we used the top-1 ranked bounding box
generated by STL directly without any extra learning process. We will use this to
compare the effect of using linear SVM classifier if trained on top-1 STL bounding
boxes as proposed in ”Method(2)”.
6.1
Performance of SVM classifier
In this part, we show some results for evaluating SVM (Method(1) in Phase I) to re-rank the
STL bounding boxes. Figure 11 shows applying SVM to re-rank top four STL bounding
boxes and then selecting the bounding box with the highest score gives much better results
on most of videos than using Top-1 STL bounding box only. The average error rate for
TOP-1 STL = 64%; while using TOP-SVM = 42%.
It should be noted there are no common ground-truth used by researchers for this case, and
the true labels are generated by manually annotating the bounding boxes. Although this
may be inaccurate and many researchers may be skeptical to this approach, we just report
these results to give an approximate overview of the effect of using SVM to re-rank the STL
bounding boxes. For accurate results, please see subsection (6.3 ’Overall performance and
quantitative comparison’).
Figure 10: Testing error rates using STL Top-1 bounding box vs using proposed SVM re-ranking
method applied on top four STL bounding box (Method(1)). Error rates are reported for each video.
6.2
Performance of TSVM classifier
In this part, we show evaluation results for TSVM (used in Phase II). Here we show
the Precision-Recall (PR) curves for each video, for the results of both TSVM applied
on Method(1) and TSVM applied on Method(2) in Figures [11-16]. Precision-Recall (PR)
curve reflects the relative proportions of positive and negative samples directly. The PR
curves are shown side-by-side for both methods in order to easily compare results of both
methods. Also, on the top of each figure, we show the Area Under the Curve (AUC) value,
which can be used to summarize the overall quality of ranking in terms of precision and
recall. The AUC is obtained by trapezoidal interpolation of the precision. An alternative
and usually almost equivalent metric is the Average Precision (AP), also shown on the
top of the figures. This is the average of the precision obtained every time a new positive
sample is recalled. It is the same as the AUC if the precision is interpolated by constant
7
segments. Additionally, we show the 11 points interpolated average precision, on the top
of each figure. This is obtained by taking the average of eleven precision values [9].
It should be noted comparing both Method(1) and Method(2) using the evaluation of
TSVM provided useless. For example, if we got low quality bounding boxes generated
in phase I, and then applied TSVM on them; regardless of how good TSVM classifier is, it
does not matter and will not improve the final results. However, if phase I generated good
bounding boxes (high quality), we then should have a good TSVM classifier to generate
final good results. In summary, having good TSVM classifier is essential but that doesn’t
guarantee good final results; as it all depends on the overall results of phase I and phase II.
In this part, we only report these TSVM performance results for showing the quality of the
classifier. Instead the overall results shown later in the ’overall results’ section, should be
used for building a reasonable comparison between both methods. Additionally, it should
be noted that the testing labels used in this evaluation are automatically generated. In this
case, the label is considered true if there is at least 40% of the ground-truth object overlaps
with the detected (test) object. Although this may not be accurate, but it can provide a
approximate measure for the performance of TSVM classifier.
In case of video ’Girl’: Figure 11 shows that TSVM applied on Method(2) performs better
than TSVM applied on Method(1) in general. In this case, Method(2) shows slightly more
successful results than Method(1). The reason behind this, is that the selected bounding
boxes (primary objects) for Method(2) are very like to each other, in both appearance
and size. Although this is good point for Method(2), but it doesn’t guarantee a better
overall performance in terms of average error pixels metric used in Segtrack dataset; that
is because this is just evaluation of performance TSVM, regardless of the quality of the
primary objects selected in Phase I.
PR (AUC: 53.95%, AP: 54.11%, AP11: 58.07%)
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
precision
precision
PR (AUC: 40.48%, AP: 41.47%, AP11: 43.75%)
1
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
PR
PR rand.
0
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
PR
PR rand.
0
1
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
Figure 11: Precision-Recall (PR) curves of video ’Girl’. On the left, PR curve of TSVM testing
results applied on Method(1). On the right, PR curve of TSVM testing results applied on
Method(2).
In case of video ’Birdfall2’:
Figure 12 shows that TSVM applied on Method(2) also performs better than TSVM applied
on Method(1) in general. In this case, Method(2) shows significant successful results than
Method(1). If we take a deep look at the final output, we will find that most of the detected
bound boxes generated by Method(2) are very large boxes, and almost fills the whole
image. That makes typically the classification task of TSVM to be classification between
whole images, which are all similar and from the same video input; resulting in good
performance as shown in the PR curve.
8
1
PR (AUC: 86.93%, AP: 87.21%, AP11: 86.78%)
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
precision
precision
PR (AUC: 52.07%, AP: 53.80%, AP11: 54.54%)
1
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
PR
PR rand.
0
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
PR
PR rand.
0
1
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
1
Figure 12: Precision-Recall (PR) curves of video ’Birdfall2’. On the left, PR curve of TSVM
testing results applied on Method(1). On the right, PR curve of TSVM testing results applied on
Method(2).
In case of video ’cheetah’: Figure 13 shows that TSVM applied on Method(1) performs
better than TSVM applied on Method(2) in general. Unlike other videos, the primary
objects in this video discovered in Phase I, are small, contained in a tight bounding boxes
and moving through the scene.
PR (AUC: 53.89%, AP: 54.87%, AP11: 62.81%)
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
precision
precision
PR (AUC: 60.70%, AP: 61.45%, AP11: 64.55%)
1
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
PR
PR rand.
PR
PR rand.
0
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
0
1
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
Figure 13: Precision-Recall (PR) curves of video ’Cheetah’. On the left, PR curve of TSVM
testing results applied on Method(1). On the right, PR curve of TSVM testing results applied on
Method(2).
In case of video ’monkeydog’: Figure 14 shows that TSVM applied on Method(2)
performs better than TSVM applied on Method(1) in general.
9
1
PR (AUC: 59.28%, AP: 59.76%, AP11: 62.14%)
PR (AUC: 50.13%, AP: 50.87%, AP11: 53.32%)
1
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
precision
precision
0.6
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
PR
PR rand.
PR
PR rand.
0
0.5
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
0
1
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
1
Figure 14: Precision-Recall (PR) curves of video ’Monkeydog’. On the left, PR curve of TSVM
testing results applied on Method(1). On the right, PR curve of TSVM testing results applied on
Method(2).
In case of video ’parachute’: Figure 15 shows that TSVM applied on Method(1) performs
slightly better than TSVM applied on Method(2) in general.
PR (AUC: 59.36%, AP: 59.57%, AP11: 61.80%)
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
precision
precision
PR (AUC: 59.99%, AP: 60.45%, AP11: 60.34%)
1
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
PR
PR rand.
0
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
PR
PR rand.
0
1
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
Figure 15: Precision-Recall (PR) curves of video ’Parachute’. On the left, PR curve of TSVM
testing results applied on Method(1). On the right, PR curve of TSVM testing results applied on
Method(2).
In case of video ’penguin’: Figure 16 shows that TSVM applied on Method(2) performs
better than TSVM applied on Method(1) in general.
10
1
PR (AUC: 59.36%, AP: 59.57%, AP11: 61.80%)
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
precision
precision
PR (AUC: 42.60%, AP: 43.39%, AP11: 48.62%)
1
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
PR
PR rand.
0
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
PR
PR rand.
0
1
0
0.1
0.2
0.3
0.4
0.5
recall
0.6
0.7
0.8
0.9
Figure 16: Precision-Recall (PR) curves of video ’Penguin’. On the left, PR curve of TSVM
testing results applied on Method(1). On the right, PR curve of TSVM testing results applied on
Method(2).
6.3
Overall performance and quantitative comparison
In Segtrack dataset [7] used in this work, the overall performance is measured using
average per-frame pixel error rate compared to the ground-truth. Apparently, lower
values of error are better. To calculate the average per-frame pixel error, we use the
following equation:
Error =
XOR(GT, f )
.
F
where f is the segmentation labeling results of the method, GT is the ground-truth labeling
of the video, and F is the number of frames in the video.
Table 1 shows the average number of error pixels per frame compared between (Method(1)
with TSVM), (Method(2) with TSVM), and other different approaches found in the
literature. In this comparison, we used the full implementation of our proposed methods
thats including details illustrated earlier in both phase I and phase II. Best values are shown
in green color. Blue values show the results of our approaches that are near from the best
values, also some of these values are much better than other approaches (but not the best).
Comparison between “Method(1) with TSVM” and “Method(1) without TSVM”
In this part, we compare the performance of Method(1) in case of using TSVM (as proposed
the earlier sections phase I + phase II), against Method(1) without using TSVM (using SVM
only in phase I). In other words, we discover the benefits of using Transductive learning
as proposed in phase II. In order to accomplish this, we used the same average number
of error pixels per frame measure, but in this case it is calculated only on the results
retrieved by SVM classifier applied on STL Top four object proposals as illustrated earlier in
a previous section. Table 2 shows the average number of error pixels per frame compared
between (Method(1) with TSVM), (Method(2) without TSVM). From the shown results, it
11
1
Video
Girl
Birdfall2
Cheetah
Monkeydog
Parachute
Penguin
Method(1)
with TSVM
2113
1259
1013
1094
1457
1890
Method(2)
with TSVM
3452
4204
2442
2254
7016
3095
Tsai et al. [7]
1304
252
1142
563
235
1705
Chockal
al. [2]
1755
454
1217
683
502
6627
et
Zhang et
al. [10]
1488
155
633
365
220
1895
Table 1: Average number of error pixels per frame between different approaches. Lower values are
better. Best values are shown in green color. Blue values show the results of our approaches that are
near from the best values, also some of these values are much better than other approaches (but not
the best).
is clear that “Method(1) with TSVM” outperforms “Method(1) with TSVM” in five out of
six cases (videos: Birdfall2, Cheetah, Cheetah, Monkeydog, Parachute, and Penguin).
Video
Girl
Birdfall2
Cheetah
Monkeydog
Parachute
Penguin
Method(1) with TSVM
2113
1259
1013
1094
1457
1890
Method(1) without TSVM
1600
2504
1136
1427
1792
2495
Table 2: Average number of error pixels per frame of “Method(1) with TSVM” against “Method(1)
without TSVM” approaches. Lower values shown in blue bold are better
Comparison between “Method(2) with TSVM” and “Method(2) without TSVM”
In this part, we compare the performance of Method(2) in case of using TSVM (as proposed
the earlier sections phase I + phase II), against Method(2) without using TSVM (using SVM
only in phase I). Also, in this case we used the average number of error pixels per frame
measure, and it is calculated only on the results retrieved by SVM classifier applied on
STL Top four object proposals taking the top-1 bounding box as ground-truth as illustrated
earlier in a previous section. Table 3 shows the average number of error pixels per frame
compared between (Method(2) with TSVM), (Method(2) without TSVM). From the shown
results, it is clear that “Method(2) with TSVM” outperforms “Method(2) with TSVM” in all
six cases (videos: Girl, Birdfall2, Cheetah, Cheetah, Monkeydog, Parachute, and Penguin).
Video
Girl
Birdfall2
Cheetah
Monkeydog
Parachute
Penguin
Method(2) with TSVM
3452
4204
2442
2254
7016
3095
Method(2) without TSVM
4849
7084
5328
2937
8532
4017
Table 3: Average number of error pixels per frame of “Method(2) with TSVM” against “Method(2)
without TSVM” approaches. Lower values shown in blue bold are better.
12
Comparison between “Method(2) without TSVM” and “Method(3) without TSVM”
In this part, we explore the efficiency of using linear SVM learned on Top-1 STL bounding
box versus Method(3) which assumes no learning and it just takes directly the top-1 STL
bounding box as the best primary object in the current frame.
We compare the performance of Method(2) without TSVM (as proposed the earlier sections
phase I), against Method(3). In other words, we discover the benefits of using linear SVM
learning on top-1 STL bounding box as proposed in phase I . We also used the same
average number of error pixels per frame measure, and it is calculated only on the results
retrieved by SVM classifier applied on STL Top four object proposals as illustrated earlier in
a previous section. Table 4 shows the average number of error pixels per frame compared
between (Method(2) without TSVM), (Method(3) without TSVM). From the shown results,
it is clear that “Method(2) without TSVM” outperforms “Method(3) without TSVM” in four
out of six cases (videos: Birdfall2, Cheetah, Cheetah, Monkeydog, and Penguin).
Video
Girl
Birdfall2
Cheetah
Monkeydog
Parachute
Penguin
Method(2) without TSVM
4849
7084
5328
2937
8532
4017
Method(3) without TSVM
4706
7982
6809
4040
6919
6311
Table 4: Average number of error pixels per frame of “Method(2) without TSVM” against
“Method(3) without TSVM”. Lower values shown in blue bold are better
6.4
Qualitative Comparison
This part provides some visualizations and analysis for the final output of both “Method(1)
with TSVM” and “Method(2) with TSVM”. Figure 17 shows the selected frames for each
method after applying SVM in phase I and TSVM in phase II. Although the improvement
that is achieved by “Method(2)” compared to applying “Method(3)” (using the top-1 STL
bounding box without learning SVM), the generated output still shows that ”Method(2)”
tends to generate bigger bounding boxes if compared to “Method(1)”. This is reasonable
because ”Method(1)” was trained on a much better ground-truth than “Method(1)” which
uses the top-a STL bounding box as a reference; which may be inaccurate bounding box in
many cases. As a result of selecting bigger bounding boxes in “Method(2)” , the average
error pixel per frame for videos dataset are greater than the error of ”Method(1)“, which
was shown earlier in Table 1.
13
Method
Method(1)
(1)
Method (2)
Method (1) with TSVM
frame t
Method (2) with TSVM
frame t'
frame t
frame t'
Figure 17: Selected output frames with the final discovered objects for “Method(1) with TSVM“
(on the left) and “Method(2) with TSVM“ (on the right).
7
Conclusions
From the experimental results shown in the previous section, it is clear that “Method (1)
with TSVM“ is the best approach compared to the others proposed in this work. It also
gives near performance to some of the state-of-the-art methods. Additionally, we showed
that using TSVM as proposed in Phase II, enhances the performance and gives better
results than using SVM only as proposed in phase I. We also showed that the learning
process conducted using SVM in phase I (whether in case of Method(1) or Method(2)),
was essential; and that enhanced the performance greatly than using only the top-1 STL
bounding box as the primary object.
References
[1] Alessandro Bergamo, Loris Bazzani, Dragomir Anguelov, and Lorenzo Torresani.
Self-taught object localization with deep networks. arXiv preprint arXiv:1409.3964, 2014.
[2] Prakash Chockalingam, Nalin Pradeep, and Stan Birchfield. Adaptive fragments-based
tracking of non-rigid objects using level sets. In IEEE International Conference on
Computer Vision (ICCV), pages 1530–1537. IEEE, 2009.
[3] Navneet Dalal and Bill Triggs. Histograms of oriented gradients for human detection.
In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR),
volume 1, pages 886–893. IEEE, 2005.
14
[4] Thorsten Joachims. Transductive inference for text classification using support vector
machines. In ICML, volume 99, pages 200–209, 1999.
[5] Yong Jae Lee, Jaechul Kim, and Kristen Grauman. Key-segments for video object
segmentation. In IEEE International Conference on Computer Vision (ICCV), pages
1995–2002. IEEE, 2011.
[6] David
Tsai,
Matthew
Flagg,
and
coherent tracking with multi-label mrf
http://cpl.cc.gatech.edu/projects/SegTrack/.
James
M.Rehg.
optimization.
BMVC,
Motion
2010.
[7] David Tsai, Matthew Flagg, Atsushi Nakazawa, and James M Rehg. Motion coherent
tracking using multi-label mrf optimization. International journal of computer vision,
100(2):190–202, 2012.
[8] A. Vedaldi and B. Fulkerson. Vlfeat: An open and portable library of computer vision
algorithm. In http://www.vlfeat.org/.
[9] Dong Zhang, Omar Javed, and Mubarak Shah. Video object segmentation through
spatially accurate and temporally dense extraction of primary object regions. In IEEE
Conference on Computer Vision and Pattern Recognition (CVPR), pages 628–635. IEEE,
2013.
15