Vision-based violence detection through deep learning
In the present society, video surveillance systems have continued to develop and incorporate more sophisticated video analysis to enhance security and public safety. With increasing demand, the need for accurate and efficient violence detection in video footage has become more critical. However, det...
Saved in:
Main Author: | |
---|---|
Format: | Final Year Project / Dissertation / Thesis |
Published: |
2024
|
Subjects: | |
Online Access: | http://eprints.utar.edu.my/6826/1/2004757_KOH_WEI_ZHE.pdf http://eprints.utar.edu.my/6826/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
my-utar-eprints.6826 |
---|---|
record_format |
eprints |
spelling |
my-utar-eprints.68262024-11-21T05:56:11Z Vision-based violence detection through deep learning Koh, Wei Zhe Q Science (General) QA75 Electronic computers. Computer science T Technology (General) In the present society, video surveillance systems have continued to develop and incorporate more sophisticated video analysis to enhance security and public safety. With increasing demand, the need for accurate and efficient violence detection in video footage has become more critical. However, detecting violence in video footage remains challenging due to varying lighting conditions and data quality. While advancements in deep learning techniques can improve the accuracy and robustness of violence detection, they often require extensive datasets, leading to overloaded training processes. This research focuses on advancing and utilizing deep learning models for violence detection in surveillance videos, with particular emphasis on varying lighting conditions. A dataset of 2,000 videos mostly in normal lighting conditions is used to train a hybrid deep learning model combining MobileNet-v2, a lightweight Convolutional Neural Network (CNN), with BiLSTM (Bidirectional Long Short-Term Memory). This hybrid model seeks to employ MobileNet-v2 for feature extraction and BiLSTM for temporal analysis in video datasets. To enhance detection accuracy under different lighting conditions, histogram equalization is integrated into the video prediction process alongside the trained base model. The approach is designed to optimize video-based violence detection without overwhelming the model with large datasets and excessive training times. The base model (MobileNet-v2 and BiLSTM) performed well in normal light conditions (96.33%). While the base model with histogram equalization achieved higher accuracy (98.91%) and the model trained on varying lighting conditions further improved to (99.15%). On the other hand, the base model performed poorly in very dark conditions (24.89%) but showed significant improvement with histogram equalization (92.21%), nearly matching the performance of the base model trained on varying lighting conditions (99.97%). This result highlights the benefit of the proposed histogram equalization method, which achieves high detection accuracy without relying on extensive datasets and overloaded training resources, making it a potential solution for real-time violence detection in diverse lighting scenarios. 2024 Final Year Project / Dissertation / Thesis NonPeerReviewed application/pdf http://eprints.utar.edu.my/6826/1/2004757_KOH_WEI_ZHE.pdf Koh, Wei Zhe (2024) Vision-based violence detection through deep learning. Final Year Project, UTAR. http://eprints.utar.edu.my/6826/ |
institution |
Universiti Tunku Abdul Rahman |
building |
UTAR Library |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Tunku Abdul Rahman |
content_source |
UTAR Institutional Repository |
url_provider |
http://eprints.utar.edu.my |
topic |
Q Science (General) QA75 Electronic computers. Computer science T Technology (General) |
spellingShingle |
Q Science (General) QA75 Electronic computers. Computer science T Technology (General) Koh, Wei Zhe Vision-based violence detection through deep learning |
description |
In the present society, video surveillance systems have continued to develop and incorporate more sophisticated video analysis to enhance security and public safety. With increasing demand, the need for accurate and efficient violence detection in video footage has become more critical. However, detecting violence in video footage remains challenging due to varying lighting conditions
and data quality. While advancements in deep learning techniques can improve the accuracy and robustness of violence detection, they often require extensive datasets, leading to overloaded training processes. This research focuses on advancing and utilizing deep learning models for violence detection in surveillance videos, with particular emphasis on varying lighting conditions. A dataset of 2,000 videos mostly in normal lighting conditions is used to train a hybrid deep learning model combining MobileNet-v2, a lightweight Convolutional Neural Network (CNN), with BiLSTM (Bidirectional Long Short-Term Memory). This hybrid model seeks to employ MobileNet-v2 for feature extraction and BiLSTM for temporal analysis in video datasets. To enhance detection accuracy under different lighting conditions, histogram equalization is integrated into the video prediction process alongside the trained base model. The approach is designed to optimize video-based violence
detection without overwhelming the model with large datasets and excessive training times. The base model (MobileNet-v2 and BiLSTM) performed well in normal light conditions (96.33%). While the base model with histogram equalization achieved higher accuracy (98.91%) and the model trained on
varying lighting conditions further improved to (99.15%). On the other hand, the base model performed poorly in very dark conditions (24.89%) but showed significant improvement with histogram equalization (92.21%), nearly matching the performance of the base model trained on varying lighting
conditions (99.97%). This result highlights the benefit of the proposed histogram equalization method, which achieves high detection accuracy without relying on extensive datasets and overloaded training resources, making it a
potential solution for real-time violence detection in diverse lighting scenarios.
|
format |
Final Year Project / Dissertation / Thesis |
author |
Koh, Wei Zhe |
author_facet |
Koh, Wei Zhe |
author_sort |
Koh, Wei Zhe |
title |
Vision-based violence detection through deep learning |
title_short |
Vision-based violence detection through deep learning |
title_full |
Vision-based violence detection through deep learning |
title_fullStr |
Vision-based violence detection through deep learning |
title_full_unstemmed |
Vision-based violence detection through deep learning |
title_sort |
vision-based violence detection through deep learning |
publishDate |
2024 |
url |
http://eprints.utar.edu.my/6826/1/2004757_KOH_WEI_ZHE.pdf http://eprints.utar.edu.my/6826/ |
_version_ |
1817849303572414464 |
score |
13.223943 |