TY - GEN
T1 - Efficient and Fast Traffic Congestion Classification Based on Video Dynamics and Deep Residual Network
AU - Abdelwahab, Mohamed A.
AU - Abdel-Nasser, Mohamed
AU - Taniguchi, Rin ichiro
N1 - Publisher Copyright:
© 2020, Springer Nature Singapore Pte Ltd.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2020
Y1 - 2020
N2 - Real-time implementation and robustness against illumination variation are two essential issues for traffic congestion classification systems, which are still challenging issues. This paper proposes an efficient automated system for traffic congestion classification based on compact image representation and deep residual networks. Specifically, the proposed system comprises three steps: video dynamics extraction, feature extraction, and classification. In the first step, we propose two approaches for modeling the dynamics of each video and produce a compact representation. In the first approach, we aggregate the optical flow in front direction, while in the second approach, we use a temporal pooling method to generate a dynamic image describing the input video. In the second step, we use a deep residual neural network to extract texture features from the compact representation of each video. In the third step, we build a classification model to discriminate between the classes of traffic congestion (low, medium, or high). We use the UCSD and NU1 traffic congestion datasets to assess the performance of the proposed method. The two datasets contain different illumination and shadow variations. The proposed method gives excellent results compared to state-of-the-art methods. It also can classify the input video in a short time (37 fps), and thus, we can use it with real-time applications.
AB - Real-time implementation and robustness against illumination variation are two essential issues for traffic congestion classification systems, which are still challenging issues. This paper proposes an efficient automated system for traffic congestion classification based on compact image representation and deep residual networks. Specifically, the proposed system comprises three steps: video dynamics extraction, feature extraction, and classification. In the first step, we propose two approaches for modeling the dynamics of each video and produce a compact representation. In the first approach, we aggregate the optical flow in front direction, while in the second approach, we use a temporal pooling method to generate a dynamic image describing the input video. In the second step, we use a deep residual neural network to extract texture features from the compact representation of each video. In the third step, we build a classification model to discriminate between the classes of traffic congestion (low, medium, or high). We use the UCSD and NU1 traffic congestion datasets to assess the performance of the proposed method. The two datasets contain different illumination and shadow variations. The proposed method gives excellent results compared to state-of-the-art methods. It also can classify the input video in a short time (37 fps), and thus, we can use it with real-time applications.
UR - http://www.scopus.com/inward/record.url?scp=85090039901&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85090039901&partnerID=8YFLogxK
U2 - 10.1007/978-981-15-4818-5_1
DO - 10.1007/978-981-15-4818-5_1
M3 - Conference contribution
AN - SCOPUS:85090039901
SN - 9789811548178
T3 - Communications in Computer and Information Science
SP - 3
EP - 17
BT - Frontiers of Computer Vision - 26th International Workshop, IW-FCV 2020, Revised Selected Papers
A2 - Ohyama, Wataru
A2 - Jung, Soon Ki
PB - Springer
T2 - International Workshop on Frontiers of Computer Vision, IW-FCV 2020
Y2 - 20 February 2020 through 22 February 2020
ER -