Download as pdf or txt
Download as pdf or txt
You are on page 1of 13

Soft Computing

https://doi.org/10.1007/s00500-020-05042-z (0123456789().,-volV)(0123456789().,-volV)

METHODOLOGIES AND APPLICATION

An intelligent multiple vehicle detection and tracking using modified


vibe algorithm and deep learning algorithm
D. Sudha1 • J. Priyadarshini1

 Springer-Verlag GmbH Germany, part of Springer Nature 2020

Abstract
Multiple vehicle detection is a promising and challenging role in intelligent transportation systems and computer vision
applications. Most existing methods detect vehicles with bounding box representation and fail to offer the location of
vehicles. However, the location information is vigorous for several real-time applications such as the motion estimation
and trajectory of vehicles moving on the road. In this paper, we propose an advanced deep learning method called enhanced
you only look once v3 and improved visual background extractor algorithms are used to detect the multi-type and multiple
vehicles in an input video. More precisely, tracking is to find the trace of the upcoming vehicles using a combined Kalman
filtering algorithm and particle filter techniques. To improve the tracking results, further, we propose the technique, namely
multiple vehicle tracking algorithms, and tested with different weather conditions such as sunny, rainy, night and fog in
input videos of 30 frames per second. The major research issues were found in the recent kinds of literature in ITS sector
which is closely related to the real-time traffic environmental problems such as occlusions, camera oscillations, background
changes, sensors, cluttering, camouflage, varying illumination changes in a day- and sunny and at nighttime vision. The
experimental results are tested with the ten different input videos and two benchmark datasets KITTI and DETRAC. The
most eight high- level features have been considered for automatic feature extraction and annotation. The attributes are
length, width, height, number of mirrors and wheels and windscreen shielding glass to detect the target region of interest
(vehicles) on road. In addition, further experiments are carried out in multiple-input videos of high definition quality using
a monocular camera, and the average accuracy is 98.6%, and the time complexity of the algorithm is O(n) and also tracking
results attained 96.6%. The dataset and input videos are discussed in comparative results with the F-test measure done for
multiple vehicles.

Keywords Multiple vehicle detection  EYOLO  IVIBE  CKFA  Occlusion

1 Introduction extensive power and resources of sensors along with


reducing the mortality rate. Among all these vehicle
Multiple vehicle detection and tracking are one of the detections and tracking is to save lives by preventing
emerging hot challenging research issues in the field of accidents on roads using the advanced technologies and
intelligent transportation sectors, computer vision, and so algorithms that exist (Chen et al. 2018). More number of
on. The motivation of vehicle detection is as follows: high existing challenges like illumination changes, camera
computational cost, lack of datasets, high error rate, oscillations, background noise, camouflage, motion speed,
occlusion and collision leads to failure detection and
tracking of multiple vehicles in real-time motion on road.
Communicated by V. Loia. Hence, various existing studies have proved that the
detection methods were obtained based on feature-based
& D. Sudha
dsudha.89@gmail.com object detection, motion-based detection, sensor-based
detection like radar, lidar and cameras and so on (Deng
J. Priyadarshini
vinodha.nizanth@gmail.com et al. 2017; Cai et al. 2017; Satzoda and Trivedi 2016; Lu
et al. 2018; Liu et al. 2010). Many machine learning and
1
School of Computer Science and Engineering, VIT, Chennai, deep learning algorithms have been proposed for object
India

123
D. Sudha, J. Priyadarshini

detection and tracking such as supervised and unsupervised The rest of the work is organized as follows: Sect. 2
machine learning, reinforcement learning, convolutional explains the pictorial representation of existing related
neural network, sliding window approach, R-CNN, fast works which has been investigated to solve the multiple
R-CNN and you only look once, multi-scale deep convo- vehicle detection and tracking. Section 3 explores the
lutional neural network. proposed method for detection and tracking methodologies
Many deep learning algorithms like fast R-CNN have to solve real-time traffic on the road. Section 4 discusses
achieved promising results for object detection, whereas a the experimental and comparative results of existing and
pretrained network is fine-tuned to classify targets based on proposed models of all vehicles in motion, Sect. 5 deals
the proper labeling with the candidate region of interest. It with the conclusion, and Sect. 6 delivers about the future
enables end-to-end training so that the processing time is scope and research directions.
slow for forward passes on every iteration of a single object
independently and sharing less computation (Gustafsson
and Frederick 2018; Ravikumar et al. 2017). To reduce the 2 Related work and issues
factors of the forward pass in R-CNN, a deep convolutional
neural network and (Chen et al. 2018) fast R-CNN have In general, three-tier categorical methods of object detec-
been introduced. Object detection is the one to locate and tion methods are considered for solving a motion detection
find instances of real-world objects like multiple cars, problem, namely background subtraction, Lu et al. (2018),
buses, trucks, bicycles based on the high-level and low- optical flow and feature-based approaches; there exists a
level features of labeled information. Moreover, object huge necessity in terms of accuracy and less computational
detection falls into three-level categories such as motion- time. In this work, we use only a monocular camera to
based object detection, feature-based object detection and capture real-time Indian traffic input videos on road. To
offline-based detection methods. Now a days, real time reduce the computational time and expensive sensors,
higher end applications lies on detection and tracking of proceed with the single monocular handheld camera of
multiple kinds of multiple vehicles on road with several high quality of 30 frames per second. There might be
challenges and deep learning plays a major role to attain several challenges for the elimination of background noise,
promising results for ITS sector problem. segmenting the objects from foreground nature along with
Deep learning has been bringing new technologies for aptly fit training models for detection algorithm with less
providing a higher level of accuracy and consumes a high error rate as depicted in Fig. 1.
computational cost of object detection with labeled pre- There has been an increasing challenge due to the
trained network model, but it is not suitable for real-time environmental factors like rainy, sunny, night, cloudy in
traffic input videos of high quality to track moving multiple nature along with occlusion, collision, size of the object,
vehicles on road. Deep learning models work well only for shape, oscillations in camera and so on. Researchers have
single object effectively than the traditional methods of AI been proposed many algorithms and methods to solve these
and ML state-of-the-art approaches in the literature. To various problems in different novels and ideas to be
overcome the existing challenges, proposed method implemented to get the promising results of such real-time
enables to attain feature based 3-tier methods Haar, LBP traffic problems to solve the higher-end level applications
and HOG and then for tracking CKFA and kalman filtering like traffic monitoring, lane estimation, and prediction,
algorithms along with modified hungarian algorithm were occlusion and collision.
used. Motivated by these real-time problems and thereby
This paper contributes to the literature in two aspects: increasing a huge number of fatality rates every year
because of the numerous challenges on road like varying
(1) We propose a novel deep learning method, enhanced
illumination changes, collision, lane following, etc. lead to
you only look once (EYOLO) v3, in order to detect
a high amount of road traffic accidents in every year. To
the multiple vehicles and represent a rectangular
solve this real-time problem, most of the scientists and
bounding box with a different classification of labels
researchers, and young researchers (Satzoda and Trivedi
like the car, bus, truck and lorry
2016) have focused toward to detect the collision and
(2) We propose a novel strategy, called combined
prevent accidents on road using a powerful sensor on toll to
filtering techniques and multiple vehicle tracking
indicate a text message globally to prevent accident and
algorithms (MVTA), to initialize window size,
reroute to other way by near 500 meters from the toll gate
correlation mean and covariance and to update the
on road (Rangesh et al. 2016). The following studies have
region proposals to find the trace of the vehicles in
been found and investigated various research experts in the
the upcoming sequences of frames.
fields of ITS to solve real-time traffic problems by deep
learning, machine learning and artificial intelligence

123
An intelligent multiple vehicle detection and tracking using modified vibe algorithm and…

Fig. 1 Illustration of average


road accidents in the Indian road
traffic

techniques and advanced algorithms, and new proposed OYOLO is faster than YOLO; detection of multiple cars
novels exist. Existing methods and techniques are listed in becomes more accurate and robust for pedestrians, cars and
Table 1. humans walking on the road. The limitation of this work is
Aleksa et al. (2018) used YOLO (Alahi et al. 2014) the lack of multiple different environments. The proposed
process to spot instantaneous vehicles on movement. The method overcomes the solutions of multi-type multiple
trained system identified categories of vehicles, namely vehicle detection of EYOLOv3 which uses squared error
car, truck, pedestrians, traffic signs and traffic lights. loss for each bounding box and class loss by changing the
Analysis of outcome showed experimental results were weights in each epoch.
working well for identifying vehicles and speedily vehicles
very near to the camera. The failure of detection occurred
for two circumstances: for heavy traffic area and when 3 Proposed work and algorithms
there was occlusion (Kim and Jeong 2014). Xu et al. (2018)
proposed a recognition scheme for vehicles which was 1. Multiple vehicle detection
captured from an aerial view with the background infor-
Our proposed method consists of 5 different layers,
mation, but due to this, the vehicle appeared small. To
namely preprocessing layer, feature extraction layer,
solve this problem, combined YOLO for dense topology
training layer, testing layer and vehicle detection layer. The
and optimal pooling was used. The experiments showed
preprocessing layer consists of more than 50,000 images of
that the model can extract features of the minor object. Lin
positive and negative images in a dataset. A positive image
et al. (2018) (Tian et al. 2014) developed vehicle counting
is the one that matches the target object of car, bus, truck
for diverse locations and directions. This developed system
and so on. The negative image is one that does not match
can give high accuracy if the video has a good lighting
with the target object of apart from vehicles like balls,
environment.
leaves, streetlights, etc. Then, feature extraction is the most
Tao et al. (2017) constructed vehicle identification for
important step for object detection in order to select the
images on road using an optimized YOLO method. In this
right features for achieving promising solutions to our
method, the last two fully connected layers are detached
problem. It is the process of extracting eight high-level and
and added an average pool layer (Li et al. 2014). The
low-level features such as length of the object, width/height
simulation results shows that the accuracy and precision
of the object, side mirrors of an object, front doom mirror
rate is higher than single object detection. Gustaffson
along with wind screening glass of each vehicle. After
(2018) (Kim and Jeong 2014) worked on linear and non-
selecting the right feature, training is needed for system
linear motions. For linear motion, Kalman filter is used,
automation using combined three-tier levels of features and
and for nonlinear motion extended Kalman filter [EKF] is
classifiers were used Haar ? AdaBoost, HOG ? SVM,
utilized (Ćorović et al. 2018; Xu et al. 2018). The EKF uses
LBP ? Fast R-CNN, while selecting the most relevant
the Jacobian matrix to linearize the nonlinear models, and
attributes in a feature selection to a new reduced set of
also tracking performance is improved. Unscented Kalman
features in feature extraction.
filter collects a set of well-known annotated samples using
a nonlinear function for calculating the mean and covari-
3.1 Preprocessing layer
ance of tracking multiple vehicles (Gustaffson 2018) (Lin
and Sun 2018).
In this preprocessing layer, a training set is created which
Single kinds of multiple vehicle tracking on road is
contains a huge collection of data for the system training to
mainly based on a set of random weights and particle filter
detect objects automatically in a given input video. Then,
(Tao et al. 2018). Ravikumar et al. (2017) developed
the system is trained with a large number of positive and
OYOLO for the target detection system on road, and

123
D. Sudha, J. Priyadarshini

Table 1 Existing methods and techniques from various literature works and proposed solutions
Author names Methods and work Benefits Limitations Solutions

Kellner et al. Sparse and low-rank Provides better performance Not suitable for multiple Extraction of multiple
(2016) Bo- constraint model by of single scenario-based scenarios moving object features on all vehicles and
Haochen contextual regularization moving object detection detection environment achieves multiple moving
et al. approach for motion vehicle detection by
detection classifiers and machine
learning approaches
Bhaskar et al. Coupled region-based Detect small objects Limited with the features and Our Proposed method
(2015) convolutional neural accurately and faster than lack of comparisons with evaluates with 7 different
Zhipeng networks for accurate multiple objects in real time the datasets of private and datasets of 3 standards and 4
Deng et al. vehicle detection public self-datasets of day, night,
sunny mode of multiple
moving vehicle detection
with tracking that exists
Hassannejad Bagging(bootstrap Yielded better accuracy and Generates target training Solutions to these problems,
et al. (2015) aggregating) mechanisms system automation is samples with confidence our methods are feature
Yingfeng used for scene adaptive missing, i.e., lack of training scores, cannot adapt with extraction by manual
Cai et al. vehicle detection algorithm and testing to the system for the automation of moving handheld and trained to the
based on a composite deep system automation to detect vehicle detection on road system for automation to
structure objects detect the 3- and 4-wheeler
objects detections in a
global manner
Battiato et al. Active learning and symmetry Yields promising accuracy for Multiple different kinds of To eliminate these existing
(2015) method which achieved the same kind of multiple vehicle detection are not problems, the proposed
Ravikumar multipart vehicle detection vehicles in terms of multiple considered method focuses on multiple
Satzoda by Haar-like features with car detection cars, bus, lorry and auto-
AdaBoost classifier for fully detection and tracking using
and partially visible rear hybrid classifiers and
view of vehicles machine learning along with
filtering techniques that
exist
Zhao et al. Generalized Haar filter-based Reduced the number of input Limited properties are Manual feature extraction
(2015) object detection for car- patches, but there are a chosen such as memory, along with the creation of
Keyu Lu sharing services by sparse better accuracy and less efficiency and lack of high- dataset and model is
et al. window techniques error rate of 6% in terms of level and low-level features proposed for front and rear
multiple patches of input and no training view of detecting and
samples mechanisms exist tracking multiple moving
vehicles on the road
Wang Single object detection using Works well only for multi- Limited target objects and Performs in multiple targets
et al.(2015) template matching from scale contrast background environments of car, bus, auto and truck
Tie Liu large database collection for single object detection along with tracking in linear
rather than multiple objects as well as nonlinear motion
also done

negative images to detect the target object and system available for a wide set of applications, the supervised
which enable it to learn the object and extract its features in learning method has been chosen as it is applicable to the
unique and derivative characteristics. Once the training is object tracking system. The system is further subjected to a
successful, it goes to the testing for accuracy of training strong classifier, namely the adaptive boost classifier
methods and results of detected objects are recorded. If it (Fig. 2).
matches with the target object, then the training is validated
and verified successfully. If not, then the training set has to
be embedded with more data in order to retrieve more
details about the object and then detect it. The artificial
intelligence technique of system learning using the Viola–
Jones method which is formally known as the Haar training
is used. Though there is a variety of learning types Fig. 2 Adaptive boost classifier

123
An intelligent multiple vehicle detection and tracking using modified vibe algorithm and…

3.2 Feature extraction The particle filter algorithm is used to track the moving
objects in both nonlinear and non-Gaussian situations. In
It is the problem of working with a large set of data in both the conditions, the moving objects in a video sequence
terms of accuracy. There exists a problem of constructing a can be tracked optimally. The steps involved are
model with a large set of data and parameters in terms of
(1) Feature extraction using SIFT
memory and computation power. When higher the
(2) Undecimated wavelet transform
parameters, there might be a computation process which is
(3) Particle filter for multiple objects tracking
also high. Feature extraction is the most prominent task in
the field of AI, machine learning, especially for classifi- 2. EYOLO v3 vehicle detection algorithm
cation of multiple objects in real-time environments.
The EYOLOv3 predicts MTMV detection using darknet
This research work mainly proceeds with the five dif-
frameworks. It involves the convolution layer, maxpooling
ferent parameters of objects length, width/height, side
layer, average connected softmax layer and output layer.
mirrors, number of tires and windscreen glass partitioning
YOLOV3 is used for predicting the bounding box and also
in the front side of the vehicle. Manual feature extraction is
calculates the objectiveness score of predicted vehicles.
done and then trained with a haar features combined with
YOLOv3 is modified to accept an extra training input of
discrete adaboost classifier forms a single strong classifier.
distance and also predict it. It uses a single convolutional
neural network to predict the bounding box together with
3.3 Object detection layer
the class probabilities. Figure 4 represents that the given
input image is split into an S 9 S grid, and within every
Once the training phase gots over then testing phase is
grid, m bounding boxes are generated. The network outputs
executed for checking the target object detection is correct
the class probability and offsets values for each of the
or not. Therefore the 3-tier object detection models were
bounding boxes, and the bounding boxes with the proba-
used for detecting the multiple vehicles in motion.
bility larger than the threshold will represent the objects
The important factor is for selecting the right features,
detected. Enhanced YOLO(EYOLO)v3 neural network is
because feature selection plays an important role in mul-
very simple and fast. It processes images in real time at 35
tiple object tracking. In general, the most desirable prop-
frames per second, and the EYOLO algorithm is also dis-
erty of a visual feature is its uniqueness so that the objects
cussed for multiple vehicle detection in real-time images or
can be easily distinguished in the feature space. Feature
videos.
selection is closely related to the object representation as
depicted in Fig. 3.

Fig. 3 Object detection engine

123
D. Sudha, J. Priyadarshini

Algorithm 1: Enhanced You Only Look Once(EYOLO)v3

Step 1: Input image, I, Detection of Interest, Output Layer

Step 2: Split I into S x S grid on the input image and then pass to the pooling layer

Step 3: The network outputs the class probability and offsets values for each of the bounding boxes

Step 4: if (class prob >confidence)

4.1 Find the bounding box with class label

4.2 Nested if( Bounding box < out of class label)

4.3 Goto if

4.4 Else if (region > bounding box)

4.5 Class= 1; //presence of vehicles

4.6 Class = 0;// absence of vehicles

4.7 End else if

4.8 End

Step 5: Plot the bounding box of finding the target region of interest

Step 6: find average mean, ROI.

Fig. 4 Enhanced YOLO v3 vehicle detection

123
An intelligent multiple vehicle detection and tracking using modified vibe algorithm and…

3. Multiple vehicle tracking The above algorithm and Fig. 5 show the process of
tracking is done for different kinds of multiple front view
Tracking is done by template matching techniques. A
vehicles such as a car, bike, truck and bus in a motion of
template is like a memory about the object. The filter bank
day-, nighttime environments with varying illumination
tracks the object using its position and velocity. It is to link
conditions. And also, we have worked with different
a sequence of object blobs together across image frames in
aspects of motion like linear and nonlinear motion; existing
order to determine the identity and location of several
approaches are modified into a new algorithm with the
objects at different points in an image sequence (blob
proof of time complexity O (log n) times. For a single
tracking). To achieve this, it is useful to describe each
iteration, it takes n times of logarithmic analysis to track in
object blob by a set of attributes, for example, its position,
motion. The updated technique is modified with various
velocity, and size, collectively called its state.
parameters and existing methods. The novelty of our pro-
Tracking the object’s true position is done by tracking
posed method and algorithm is for parameter tuned with
its state using a Kalman filter bank. This uses information
different multi-modalities like vehicle length, width/height,
from the current blob and the previous object state to create
side mirror, number of tires, front wind screening glass and
an estimate of the object’s new state. The combination of
shape which leads to accurate training model than the
previously estimated object dynamics and current mea-
existing state of the art.
surements helps eliminate noise from blob measurements
In step 4, each iteration exists with the update of each
that would otherwise lead to erratic object tracking.
target of its position and velocity along with the direction
Algorithm 2 discusses the previous position and velocity of
of lane following so that our method improves the overall
a vehicle that allows us to give a rough estimate of its
accuracy rate is 5 times of average than the existing art of
current position. When combining this estimate with
82.6%.
additional information of its state, tracking accuracy can be
overall improved.

123
D. Sudha, J. Priyadarshini

Fig. 5 Multiple vehicle tracking layer using RPN

4 Experimental analysis are decomposed into vectors combining the x-axis and
y-axis components.
To verify the proposed method, various experiments were
carried out on both continuous data and real images. Fig- 4.2 Test and result comparison
ure 6 depicts the error plots of using the Kalman filter. A
graph is plotted for both the real observation of the sine Experimental results were performed on a high-speed Intel
wave and the estimated observation. The observations are Core i5 processor of a standard desktop computer using
predicted and estimated using the linear Kalman filter with MATLAB (R2016a). A total of 2,46,345 images have been
time in the X-axis and amplitude in the Y-axis. In the linear used for training and testing. All these images were col-
Kalman filter, the estimated observation is almost similar lected on 10 different input videos of capturing on the road
to the real observation. It predicts the behavior of moving of day and night. We have a collection of 10 input videos
objects more accurately than the EKF and UKF. But it of day and night moving vehicles of car, bike, bus, lorry,
cannot be applied to real-world moving objects since their etc. In our proposed algorithm, multiple vehicle tracking
process equation tends to be nonlinear. has been done for bike, car, bus and lorry in a day- and
nighttime dataset and videos. Table 2 represents the com-
4.1 Data collection putation time calculation of different datasets with the
efficiency of detection and tracking results that exist. The
The data were collected and then used to validate the graph shows the computational time complexity of speed
simulation model of collecting day and night moving and velocity of each vehicle, respectively. Various graphs
vehicles. Input videos are collected between 10–12 am and were plotted for efficiency Vs calculating the speed of
6–7 pm on Indian highways and traffic roads. The private tracking and detecting, respectively.
dataset provides the collection of positive and negative Multiple vehicle detection using enhanced YOLO v3 for
images of.jpeg images of equal size; then, there is a chance the sunny and rainy dataset is tabulated in Table 2.
to get the respective input videos of vehicle speed, motion, The experimental result using MYOLOv3 for detection
and acceleration and input size of trajectories length and tested on the night and fog dataset is tabulated in Table 3,
velocity. Foreground detection is done by using the and Fig. 6 shows the F-test comparison of estimated versus
removal of background noise using a background sub- actual result attained in the statistical test with 10 different
traction method. Finally, velocity and acceleration values input videos of 30 frames/s.

123
An intelligent multiple vehicle detection and tracking using modified vibe algorithm and…

Table 3 Experimental results of multiple vehicle detection using


EYOLO
Datasets Average accuracy (%) Precision Recall

Night dataset 97.5 0.80 0.74


Fog detection 96.9 0.87 0.72

lorry in motion as a total count of moving vehicles as


mentioned in density in a sunny environment of daytime
motion. Figure 8 represents the multiple vehicle detection
of tracking vehicles, respectively; for example, the car is
bounded with black color rectangle representation and
other objects are marked with orange color and bus is
marked with green color and in linear motion on road.
Computational time analysis is tabulated in Tables 4 and
5. To calculate the speed analysis of different resolutions
for Multiple object tracking of object size and speed of
different video resolution of datasets and F-test score is
computed as shown in Table 6.
The speed of the proposed model is analyzed by sub-
jecting the system to different datasets where the object
size was variable in nature. On average, it was found that
the speed of the tracking (Table 6) in the case of the single
object was found to be as high as 13 fps. This is an
approximately 46% increase in speed in comparison with
the approach proposed (Tian et al. 2014).
Figure 8 represents that the multiple vehicle detection in
the front view of a standard dataset of benchmark KITTI
found on the results of proposed model achieves 92.6% in
Fig. 9 true positive rate than the existing state of the art
which has been discussed in a literature survey of various
works related to multiple detection and tracking.
Comparing existing multiple hypothesis tracking algo-
rithms with MVTA, the accuracy, precision and recall are
tabulated in Table 7.
Figure 10 represents the overall performance analysis of
Fig. 6 Mean error plot using Kalman and unscented Kalman filter multi-type and multiple vehicle(MTMV) detection and
bank tracking, and the accuracy is 97.69% improved 5% of
greater accuracy than existing techniques.
Table 2 Experimental result of multiple vehicle detection using
IVIBE
Dataset Accuracy (%) Precision Recall 5 Conclusion
Sunny dataset 98.5 0.82 0.86
The inference of the proposed work is concluded with the
Rainy dataset 96.7 0.78 0.68
different vehicle detection and tracking which clearly
estimates the information about the object region with the
help of machine learning, collection of the system object
model and attained a different dataset of multiple vehicle
Figure 7 represents the results of multiple vehicle detection and tracking by particle and extended Kalman
detection in private datasets 1 and 2 compared with the filtering techniques. The primary objective of this research
benchmark dataset of DETRAC of multiple cars, bus and work is to detect and track MTMV in motion and also to

123
D. Sudha, J. Priyadarshini

Fig. 7 Multiple vehicle detection in daytime motion in private datasets 1 and 2

Fig. 8 Multiple vehicle tracking of daytime video in linear motion

Table 4 Performance analysis of multiple vehicle detection methods Table 5 Average computational time for existing filters and proposed
with benchmark datasets CKFA (in seconds)
Techniques Accuracy (%) Precision Recall Kalman filter Particle filter Proposed CKFA

IVIBE 97.6 0.82 0.76 3.25 2.86 1.84


MYOLOv3 98.9 0.88 0.78
KITTI 92.2 0.76 0.74
DETRAC 94.1 0.81 0.69
Table 6 Speed analysis for multiple object tracking
Dataset Video resolution Object size Speed (fps) F-test

increase the accuracy, precision and recall rates using I 420 9 240 282 9 92 9–11 0.0076
various methodologies. The main contribution of this work II 640 9 480 130 9 105 13–15 0.0082
is as follows: The important contribution is the detection of III 640 9 480 120 9 55 12–15 0.0078
MTMV in motion using IVIBE, MYOLO and finding the
MTMV in a sequence of frames. Using proposed

123
An intelligent multiple vehicle detection and tracking using modified vibe algorithm and…

Fig. 9 Multiple vehicle


detection in sunny video

123
D. Sudha, J. Priyadarshini

Fig. 9 continued

Table 7 Comparative results of


Technique Accuracy (%) Precision Recall
multiple vehicle tracking using
CKFA Multiple hypothesis tracking algorithm 92.7 0.80 0.72
Proposed MVTA 95.6 0.92 0.82

Fig. 10 Overall performance


analysis of MTMV detection
and tracking

algorithms, the improved accuracy is 96.8% than the sunny, rainy, night and fog. The benchmark datasets KITTI
existing method. The next contribution is tracking MTMV and DETRAC were taken from the source and for com-
using CKFA and MVTA, and the accuracy is 95.6% parisons made with the private and public datasets and
achieved than existing methods. The private dataset is results are shown in experimental results section. The
collected 2-h real-time videos in highways using a high- proposed method yields a higher accuracy rate and reduces
resolution camera and tested with 10 different private the minimal error rate shown in the performance analysis
datasets including different weather conditions such as of MTMV detection and tracking.

123
An intelligent multiple vehicle detection and tracking using modified vibe algorithm and…

Data availability We used our own data multiple model particle filter. IEEE Trans Intell Transp Syst
15(6):2480–2490
Li S, Yu H, Zhang J, Yang K, Bin R (2014) Video-based traffic data
Compliance with ethical standards collection system for multiple vehicle types. IET Intell Transp
Syst 8(2):164–174
Conflict of interest All authors state that there is no conflict of J Lin, M Sun (2018) A YOLO-based traffic counting system. In: IEEE
interest. transactions on pattern analysis and artificial intelligence,
pp 234–235
Human and animal rights Humans/animals are not involved in this Liu T, Yuan Z, Sun J, Wang J, Zheng N, Tang X, Shum HY (2010)
work Learning to detect a salient object. IEEE Trans Pattern Anal
Mach Intell 33(2):353–367
Lu K, Li J, Zhou L, Hu X, An X, He H (2018a) Generalized haar
filter-based object detection for car sharing services. IEEE Trans
References Automat Sci Eng 15(4):1448–1458
Lu K, Li J, Zhou L, Hu X, An X, He H (2018b) Generalized haar
Alahi A, Bierlaire M, Vandergheynst P (2014) Robust real-time filter-based object detection for car sharing services. IEEE Trans
pedestrian detection in an urban environment with low-resolu- Autom Sci Eng 15(4):1448–1458
tion cameras. Transp Res Part C 39:113–128 Rangesh A, Ohn-Bar E, Trivedi MM (2016) Long-term multi-cue
Battiato S, Farinella GM, Furnari A, Puglisi G, Snijders A, Spiekstra J tracking of hands in vehicles. IEEE Trans Intell Transp Syst
(2015) An integrated system for vehicle tracking and classifica- 17(5):1483–1492
tion. Expert Syst Appl 42(21):7275 Ravikumar, Ling-Feng Shi, and Xiao Ke (2017) A robust moving
Bhaskar H, Dwivedi K, Dogra DP, Al-Mualla M, Mihaylova L (2015) object detection in multi-scenario big data for video surveillance.
Autonomous detection and tracking under illumination changes, IEEE Transactions on circuits and systems for video technology,
occlusions and moving camera. Signal Process 117:343–354 2017, pp 167–172
Cai Y, Wang H, Zheng Z, Sun X (2017) Scene-adaptive vehicle Satzoda RK, Trivedi MM (2016) Multipart vehicle detection using
detection algorithm based on a composite deep structure. IEEE symmetry-derived analysis and active learning. IEEE Trans
Access 5:22804–22811 Intell Transport Syst 17(4):926–937
Chen BH, Shi LF, Ke X (2018) A robust moving object detection in Tao J, Wang H, Zhang X, Li X, Yang H (2018) An object detection
multi-scenario big data for video surveillance. IEEE Trans system based on YOLO in traffic scene. In: 6th International
Circuits Syst Video Technol 29(4):982–995 conference on computer science and network technology,
Ćorović A, Ilić V, Durić S, Marijan M, Pavković B (2018) The real- pp 315–319
time detection of traffic participants using YOLO algorithm. In: Tian B, Li Y, Li B, Wen D (2014) Rear-view vehicle detection and
IEEE access: 26th telecommunications forum (TELFOR), 2018, tracking by combining multiple parts for complex urban
pp 1–4 surveillance. IEEE Trans Intell Transp Syst 15(2):597–606
Deng Z, Sun H, Zhou S, Zhao J, Zou H (2017) Toward fast and Wang CC, Thorpe C, Suppe A (2015) Ladder based detection and
accurate vehicle detection in aerial images using coupled region- tracking of moving objects from a ground vehicle at high speeds.
based convolutional neural networks. IEEE J Selected Topics In: IEEE, pp 416–421
Appl Earth Obser Remote Sens 10(8):3652–3664 Xu Z, Shi H, Li N, Xiang C (2018)Vehicle detection under UAV
Gustafsson and Frederick (2018) Statistical sensor fusion of vehicle based on optimal dense YOLO method. In: IEEE 5th interna-
tracking on road. IEEE Access: ISBN:978-9144127248, tional conference on systems and informatics, pp 247–258
pp 218–228 Zhao N, Xia Y, Xu C, Shi X, Liu Y (2015) APPOS: an adaptive
Hassannejad H, Medici P, Cardarelli E, Cerri P (2015) Detection of partial occlusion segmentation method for multiple vehicles
moving objects in roundabouts based on a monocular system. tracking. J Vis Commun Image R 37:25–31
Expert Syst Appl 42(9):4167–4176
Kellner D, Barjenbruch M, Klappstein J, Dickmann J, Dietmayer K Publisher’s Note Springer Nature remains neutral with regard to
(2016) Tracking of extended objects with high-resolution jurisdictional claims in published maps and institutional affiliations.
Doppler radar. IEEE Trans Intell Transp Syst 17(5):1493–1509
Kim T, Jeong HY (2014) A novel algorithm for crash detection under
general road scenes using crash probabilities and an interactive

123

You might also like