Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs
Videos captured using cameras from unmanned aerial vehicles (UAV) normally produce dynamic footage that commonly contains unstable camera motion with multiple moving objects. These objects are sometimes occluded by vegetation or even other objects, which presents a challenging environment for...
Saved in:
Main Author: | |
---|---|
Format: | Thesis |
Language: | English |
Published: |
2017
|
Online Access: | http://psasir.upm.edu.my/id/eprint/68539/1/FK%202018%2024%20-%20IR.pdf http://psasir.upm.edu.my/id/eprint/68539/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
my.upm.eprints.68539 |
---|---|
record_format |
eprints |
institution |
Universiti Putra Malaysia |
building |
UPM Library |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Putra Malaysia |
content_source |
UPM Institutional Repository |
url_provider |
http://psasir.upm.edu.my/ |
language |
English |
description |
Videos captured using cameras from unmanned aerial vehicles (UAV) normally
produce dynamic footage that commonly contains unstable camera motion with
multiple moving objects. These objects are sometimes occluded by vegetation or even
other objects, which presents a challenging environment for higher level video
processing and analysis. This thesis deals with the topic of moving object detection
(MOD) whose intention is to identify and detect single or multiple moving objects
from video. In the past, MOD was mainly tackled using image registration, which
discovers correspondences between consecutive frames using pair-wise grayscale
spatial visual appearance matching under rigid and affine transformations. However,
traditional image registration is unsuitable for UAV captured videos since distancebased
grayscale similarity fails to cater for the dynamic spatio-temporal differences of
moving objects. Registration is also ineffective when dealing with object occlusion.
This thesis therefore proposes a framework to address these issues through a two-step
approach involving region matching and region labeling. Specifically, the objectives
of this thesis are (i) to develop an image registration technique based on multigraph
matching, (ii) to detect occluded objects through exploration of candidate object
correspondences in longer frame sequences, and (iii) to develop a robust graph
coloring algorithm for multiple moving object detection under different
transformations.
In general, each frame of the footage will firstly be segmented into superpixel regions
where appearance and geometrical features are calculated. Trajectory information is
also considered across multiple frames taking into account many types of
transformations. Specifically, each frame is modeled/represented as a regional
adjacency graph (RAG). Then, instead of pair-wise spatial matching as with image registration, correspondences between video frames are discovered through
multigraph matching of robust spatio-temporal features of each region. Since more
than two frames are considered at one time, this step is able to discover better region
correspondences as well as caters for object(s) occlusion. The second step of region
labeling relies on the assumption that background and foreground moving objects
exhibit different motions properties when in a sequence. Therefore, their spatial
difference is expected to drastically differ over time. Banking on this, region labeling
assigns the labels of either background or foreground region based on a proposed
graph coloring algorithm, which considers trajectory-based features. Overall, the
framework consisting of these two steps is termed as Motion Differences of Matched
Region-based Features (MDMRBF). MDMRBF has been evaluated against two
datasets namely the (i) Defense Advanced Research Projects Agency (DARPA) Video
Verification of Identity (VIVID) dataset and (ii) two self-captured videos using a
mounted camera on a UAV. Precision and recall are used as the criteria to
quantitatively evaluate and validate overall MOD performance. Furthermore, both are
computed with respect to the ground-truth data which are manually annotated for the
video sequences. The proposed framework has also been compared with existing stateof-
the-art detection algorithms. Experimental results show that MDMRBF
outperforms these algorithms with precision and recall being 94% and 89%,
respectively. These results can be attributed to the integration of appearance and
geometrical constraints for region matching using the multigraph structure. Moreover,
the consideration of longer trajectories on multiple frames and taking into account all
the transformations also facilitated in resolving occlusion. With regards to time, the
proposed approach could detect moving objects within one minute for a 30-second
sequence, which means that it is efficient in practice. In conclusion, the multiple
moving object detection technique proposed in this study is robust to unknown
transformations, with significant improvements in overall precision and recall
compared to existing methods. The proposed algorithm is designed in order to tackle
many limitations of the existing algorithms such as handle inevitable occlusions,
model different motions from multiple moving objects, and consider the spatial
information. |
format |
Thesis |
author |
Harandi, Bahareh Kalantar Ghorashi |
spellingShingle |
Harandi, Bahareh Kalantar Ghorashi Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs |
author_facet |
Harandi, Bahareh Kalantar Ghorashi |
author_sort |
Harandi, Bahareh Kalantar Ghorashi |
title |
Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs |
title_short |
Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs |
title_full |
Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs |
title_fullStr |
Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs |
title_full_unstemmed |
Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs |
title_sort |
moving objects detection from uav captured videos using trajectories of matched regional adjacency graphs |
publishDate |
2017 |
url |
http://psasir.upm.edu.my/id/eprint/68539/1/FK%202018%2024%20-%20IR.pdf http://psasir.upm.edu.my/id/eprint/68539/ |
_version_ |
1643839231062179840 |
spelling |
my.upm.eprints.685392019-05-15T00:24:07Z http://psasir.upm.edu.my/id/eprint/68539/ Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs Harandi, Bahareh Kalantar Ghorashi Videos captured using cameras from unmanned aerial vehicles (UAV) normally produce dynamic footage that commonly contains unstable camera motion with multiple moving objects. These objects are sometimes occluded by vegetation or even other objects, which presents a challenging environment for higher level video processing and analysis. This thesis deals with the topic of moving object detection (MOD) whose intention is to identify and detect single or multiple moving objects from video. In the past, MOD was mainly tackled using image registration, which discovers correspondences between consecutive frames using pair-wise grayscale spatial visual appearance matching under rigid and affine transformations. However, traditional image registration is unsuitable for UAV captured videos since distancebased grayscale similarity fails to cater for the dynamic spatio-temporal differences of moving objects. Registration is also ineffective when dealing with object occlusion. This thesis therefore proposes a framework to address these issues through a two-step approach involving region matching and region labeling. Specifically, the objectives of this thesis are (i) to develop an image registration technique based on multigraph matching, (ii) to detect occluded objects through exploration of candidate object correspondences in longer frame sequences, and (iii) to develop a robust graph coloring algorithm for multiple moving object detection under different transformations. In general, each frame of the footage will firstly be segmented into superpixel regions where appearance and geometrical features are calculated. Trajectory information is also considered across multiple frames taking into account many types of transformations. Specifically, each frame is modeled/represented as a regional adjacency graph (RAG). Then, instead of pair-wise spatial matching as with image registration, correspondences between video frames are discovered through multigraph matching of robust spatio-temporal features of each region. Since more than two frames are considered at one time, this step is able to discover better region correspondences as well as caters for object(s) occlusion. The second step of region labeling relies on the assumption that background and foreground moving objects exhibit different motions properties when in a sequence. Therefore, their spatial difference is expected to drastically differ over time. Banking on this, region labeling assigns the labels of either background or foreground region based on a proposed graph coloring algorithm, which considers trajectory-based features. Overall, the framework consisting of these two steps is termed as Motion Differences of Matched Region-based Features (MDMRBF). MDMRBF has been evaluated against two datasets namely the (i) Defense Advanced Research Projects Agency (DARPA) Video Verification of Identity (VIVID) dataset and (ii) two self-captured videos using a mounted camera on a UAV. Precision and recall are used as the criteria to quantitatively evaluate and validate overall MOD performance. Furthermore, both are computed with respect to the ground-truth data which are manually annotated for the video sequences. The proposed framework has also been compared with existing stateof- the-art detection algorithms. Experimental results show that MDMRBF outperforms these algorithms with precision and recall being 94% and 89%, respectively. These results can be attributed to the integration of appearance and geometrical constraints for region matching using the multigraph structure. Moreover, the consideration of longer trajectories on multiple frames and taking into account all the transformations also facilitated in resolving occlusion. With regards to time, the proposed approach could detect moving objects within one minute for a 30-second sequence, which means that it is efficient in practice. In conclusion, the multiple moving object detection technique proposed in this study is robust to unknown transformations, with significant improvements in overall precision and recall compared to existing methods. The proposed algorithm is designed in order to tackle many limitations of the existing algorithms such as handle inevitable occlusions, model different motions from multiple moving objects, and consider the spatial information. 2017-12 Thesis NonPeerReviewed text en http://psasir.upm.edu.my/id/eprint/68539/1/FK%202018%2024%20-%20IR.pdf Harandi, Bahareh Kalantar Ghorashi (2017) Moving objects detection from UAV captured videos using trajectories of matched regional adjacency graphs. PhD thesis, Universiti Putra Malaysia. |
score |
13.213113 |