Paper 15032
Paper 15032
IJARSCT
International Journal of Advanced Research in Science, Communication and Technology (IJARSCT)
International Open-Access, Double-Blind, Peer-Reviewed, Refereed, Multidisciplinary Online Journal
Impact Factor: 7.53 Volume 4, Issue 1, January 2024
Abstract: Traffic analysis is a problem that city planners have been dealing with for years. Smarter
methods are developed to analyze traffic and speed up the process. Traffic analysis can record the number
of vehicles and vehicle classes in an area at a given time. People have been developing such mechanisms
for decades now, but most of them involve using sensors to calculate the direction of moving vehicles and
identify vehicles to track vehicle numbers. Although this system has matured over time and is very effective,
they are not budget-friendly. The problem is that such systems require periodic maintenance and
calibration. Therefore, this project aims to calculate and classify the vehicle based on vision. The system
involves capturing frames from video to detect and count vehicles using Gaussian Mixture Model (GMM)
background subtraction, then classifying vehicles by comparing contour areas with predicted values. A
significant contribution of the paper is the comparison of two classification methods. Classification is done
using Contour Comparison (CC) and Bag of Features (BoF) methods.
I. INTRODUCTION
Today, countries and governments need a safe and affordable system to automate vehicles and control vehicle theft.
Increasing traffic on roads and highways, increasing congestion, and problems with existing vehicle detectors have led
to the development of new vehicle detection technologies. Computer vision systems are the most common choice, but
several problems must be overcome to successfully perform classification. Real-time detection and tracking of objects
or vehicles moving on different roads by intelligent vision systems is important for many fields of research and
technology applications.
Extracting useful information such as traffic density, object speed, driver behavior and traffic patterns from these
camera systems becomes essential. Manual analysis is no more. Developing intelligent systems that can extract traffic
congestion and vehicle classification data from traffic control systems is essential for traffic management. Otherwise,
the monitoring system is also important in driver assistance applications, as the vision system allows the detection and
classification of the vehicles involved in the photographed incident.
A figure is a visual representation of something. The term has several uses in information technology. A photo is an
image that is created or copied and stored electronically. An image can be interpreted as a vector graphic or a raster
graphic. Digital image processing involves the manipulation of digital images using a digital computer. This is part of
the signal and system, but pay special attention to the image. DIP focuses on developing computer systems that can
perform image processing. The input of the system is a digital image and the system processes the image using an
efficient algorithm with the image as output. It makes it possible to apply more extensive algorithms to the input image
and can avoid problems such as noise and signal distortion during processing. images can be divided into the following
three types.
A binary image consists of pixels that can be one of two colors, usually black and white. Binary images are called two
levels or two levels. This means that each pixel is stored as an integer, ie 0 or 1. Gray is an intermediate color between
white and white. It is a neutral color or achromatic color, which literally means "colorless" color because it can be
composed of white and white. It was the color of cloudy skies, dust and lead. A color (digital) image is a digital image
that contains color information for each pixel. This process is environmentally friendly as it does not require chemical
processing. Digital imaging is often used to document and record historical, scientific, and personal events
Copyright to IJARSCT DOI: 10.48175/IJARSCT-15032 207
www.ijarsct.co.in
ISSN (Online) 2581-9429
IJARSCT
International Journal of Advanced Research in Science, Communication and Technology (IJARSCT)
International Open-Access, Double-Blind, Peer-Reviewed, Refereed, Multidisciplinary Online Journal
Impact Factor: 7.53 Volume 4, Issue 1, January 2024
This paper describes a vision-based system for detecting, tracking and classifying moving vehicles. Four different
potential traffic groups can be defined, but the proposed software is flexible and the number of groups that can be
classified.
Tituana et al. review various previous works developed in this field and identify the technological methods and tools
used in those works; In addition, this study also highlights trends in this field. The most relevant articles are reviewed
and the results are summarized in tables and figures
The above equations often produce complex functions that are difficult to scale. What we can do in this case is to use
Jensens Inequality to construct a lower bound function that is easier to optimize. If we optimize this by minimizing the
KL difference (gap) between the two distributions, we can approximate the original function. This process is described.
I have also provided a link to a video showing the derivation of the KL divergence for those who want a more rigorous
mathematical explanation.
In fact, we only need to do two steps to evaluate our model. In the first step (E-step), we want to estimate the posterior
distribution of our latent variable in terms of conditional weight (π), our term (µ) and the Gaussian mean covariance
(Σ). Then we can enter the second step (M-step) and use it
Increase the likelihood associated with our parameter parameter θ. This process is repeated until the algorithm
converges (the loss function remains unchanged).
Fig. 2: An example of taking a blob of text and converting it into a word histogram.
Given a dictionary of possible visual words, we can then count the number of times each visual word appears and
visualize it as a histogram. This histogram is a veritable bag of visual words. Building visual vocabulary can be divided
into three steps
Classification
One of the interesting features of our network is its simplicity: the classifier is only replaced by a masking layer,
without any prior or convolutional structure. However, it should be prepared with a large amount of training data:
vehicles of different sizes should appear almost everywhere.
Visual tracking solves the problem of finding a target in a new frame from the current position. The proposed tracker
dynamically tracks the target with sequence movement controlled by GMM. GMM predicts the motion to run the target
from its position in the previous frame. The intersection box moves with the movement predicted from the previous
state, and the next movement continues to be predicted from the moved state. We solve the vehicle tracking problem by
repeating this process in a series of tests. GMM excels in RL as well as SL. Online adaptation takes place during real
tracking.
A GMM is designed to generate motion to find the location and size of the target vehicle in a new frame. The GMM
algorithm learns a policy that selects the most optimal action to follow from the current situation. In GMM, a policy
system is developed in which the input is a truncated image layer in the previous state and the output is the probability
distribution of actions such as translation and scale change. The process of choosing this course of action requires a bit
of research step more than the sliding window or candidate sampling approach. Furthermore, since our method can
localize the target by selecting the motion, post-processing such as box regression is not required
REFERENCES
[1] Alpatov, Boris & Babayan, Pavel & Ershov, Maksim. (2018). Vehicle detection and counting system for
real-time traffic surveillance. 1-4. 10.1109/MECO.2018.8406017.
[2] Song, H., Liang, H., Li, H. et al. Vision-based vehicle detection and counting system using deep learning in highway
scenes. Eur. Transp. Res. Rev. 11, 51 (2019). https://doi.org/10.1186/s12544-019-0390-4.
[3] Neupane, Bipul et al. “Real-Time Vehicle Classification and Tracking Using a Transfer Learning-Improved Deep
Learning Network.” Sensors (Basel, Switzerland) vol. 22,10 3813. 18 May. 2022, doi:10.3390/s22103813.
[4] C. J Lin, Shiou-Yun Jeng, Hong-Wei Lioa, "A Real-Time Vehicle Counting, Speed Estimation, and Classification
System Based on Virtual Detection Zone and YOLO", Mathematical Problems in Engineering, vol. 2021, Article ID
1577614, 10 pages, 2021. https://doi.org/10.1155/2021/1577614.
[5] M. S. Chauhan, A. Singh, M. Khemka, A. Prateek, and Rijurekha Sen. 2019. Embedded CNN based vehicle
classification and counting in non-laned road traffic. In Proceedings of the Tenth International Conference on
Information and Communication Technologies and Development (ICTD '19). Association for Computing
Machinery, New York, NY, USA, Article 5, 1–11. https://doi.org/10.1145/3287098.3287118.
[6] A. Arinaldi, J. A. Pradana, A. A. Gurusinga, “Detection and classification of vehicles for traffic video analytics”,
Procedia Computer Science, Volume 144, 2018, Pages 259-268, ISSN 1877-0509,
https://doi.org/10.1016/j.procs.2018.10.527.
[7] Gomaa, A., Minematsu, T., Abdelwahab, M.M. et al. Faster CNN-based vehicle detection and counting strategy for
fixed camera scenes. Multimed Tools Appl 81, 25443–25471 (2022). https://doi.org/10.1007/s11042-022-12370-9.
[8] G. Oltean, C. Florea, R. Orghidan and V. Oltean, "Towards Real Time Vehicle Counting using YOLO-Tiny and Fast
Motion Estimation," 2019 IEEE 25th International Symposium for Design and Technology in Electronic Packaging
(SIITME), 2019, pp. 240-243, doi: 10.1109/SIITME47687.2019.8990708.
[9] L. C. Pico and D. S. Benítez, "A Low-Cost Real-Time Embedded Vehicle Counting and Classification System for
Traffic Management Applications," 2018 IEEE Colombian Conference on Communications and Computing
(COLCOM), 2018, pp. 1-6, doi: 10.1109/ColComCon.2018.8466734.
[10] D. E. V. Tituana, S. G. Yoo and R. O. Andrade, "Vehicle Counting using Computer Vision: A Survey," 2022 IEEE
7th International conference for Convergence in Technology (I2CT), 2022, pp. 1-8, doi:
10.1109/I2CT54291.2022.9824432.
[11] A. Khan, A., Sabeenian, R.S., Janani, A.S., Akash, P. (2022). Vehicle Classification and Counting from Surveillance
Camera Using Computer Vision. In: Suma, V., Baig, Z., K. Shanmugam, S., Lorenz, P. (eds) Inventive Systems and
Copyright to IJARSCT DOI: 10.48175/IJARSCT-15032 214
www.ijarsct.co.in
ISSN (Online) 2581-9429
IJARSCT
International Journal of Advanced Research in Science, Communication and Technology (IJARSCT)
International Open-Access, Double-Blind, Peer-Reviewed, Refereed, Multidisciplinary Online Journal
Impact Factor: 7.53 Volume 4, Issue 1, January 2024
Control. Lecture Notes in Networks and Systems, vol 436. Springer, Singapore. https://doi.org/10.1007/978-981-19-
1012-8_31.
[12] W. Balid, H. Tafish and H. H. Refai, "Intelligent Vehicle Counting and Classification Sensor for Real-Time Traffic
Surveillance," in IEEE Transactions on Intelligent Transportation Systems, vol. 19, no. 6, pp. 1784-1794, June 2018, doi:
10.1109/TITS.2017.2741507.
[13] N. Jahan, S. Islam and M. F. A. Foysal, "Real-Time Vehicle Classification Using CNN," 2020 11th International
Conference on Computing, Communication and Networking Technologies (ICCCNT), 2020, pp. 1-6, doi:
10.1109/ICCCNT49239.2020.9225623.
[14] M. A. Butt, A. M. Khattak, S. Shafique, B. Hayat, S. Abid, Ki-Il Kim, M. W. Ayub, A. Sajid, A. Adnan,
"Convolutional Neural Network Based Vehicle Classification in Adverse Illuminous Conditions for Intelligent
Transportation Systems", Complexity, vol. 2021, Article ID 6644861, 11 pages, 2021.
https://doi.org/10.1155/2021/6644861.
[15] P. Gonzalez, Raul & Nuño-Maganda, Marco Aurelio. (2014). Computer vision based real-time vehicle tracking
and classification system. Midwest Symposium on Circuits and Systems. 679-682. 10.1109/MWSCAS.2014.6908506.