Robust hybrid technique for moving object detection and tracking using cartoon features and fast PCP
S.H. Jeevith1, S. Lakshmikanth 2 1 Sri Siddhartha Institute of Technology, Tumakuru-572105, India;
2 Acharya Institute of Technology, Bengaluru-560107, India
Abstract
In various computer vision applications, the moving object detection is an essential step. Principal Component Analysis (PCA) techniques are often used for this purpose. However, the performance of this method is degraded by camera shake, hidden moving objects, dynamic background scenes, and / or fluctuating exposure. Robust Principal Component Analysis (RPCA) is a useful approach for reducing stationary background noise as it can recover low rank matrices. That is, moving object is formed by the low power models and the static background of RPCA. This paper proposes a simple alternative minimization algorithm to fix minor discrepancies in the original Principal Component Pursuit (PCP) or RPCA function. A novel hybrid method of cartoon texture features used as a data matrix for RPCA taking into account low-ranking and rare matrix is presented. A new non-convex function is proposed to better control the low-range properties of the video background. Simulation results demonstrate that the proposed algorithm is capable of giving consistent random estimates and can indeed improve the accuracy of object recognition in comparison with existing methods.
Keywords: principal component pursuit, robust principal component analysis, cartoon features, local binary patterns.
Citation: Jeevith SH, Lakshmikanth S. Robust hybrid technique for moving object detection and tracking using cartoon features and fast PCP. Computer Optics 2022; 46(5): 783-789. DOI: I0.18287/2412-6179-C0-I056.
Introduction
Motion detection is the primary step in computer vision, focusing at grouping moving objects into video images. It has originated many applications for human observation, image capture, motion behavior analysis and vision-based interaction of computer and humans. Several articles on motion detection have been published over the past 20 years. However, there are still significant difficulties in detecting fast moving objects, for example, against a complex background. For example, rapidly changing the frequency of bright objects in the background (such as branches and leaves), camera shake, side effects, stationary foreground, efficiency, etc. It is important to note that detecting static images and background is a trade-off as anyone expects to remove as many backgrounds as favorable while maintaining the forefront. Though, it is challenging to achieve the desired outcome because they have many similar functions. Existing methods rarely solve this problem [1]. Conventional motion detection techniques are established on noise reduction, which usually involves three phases. First, the background image is mathematically defined. Second, new observations of the background or foreground are ranked allowing to the resemblance between the background and observations. At final phase, updating the background model for new observations using an adaptive strategy.
Liu and Zhang [2] utilized the difference in histogram resemblance between the imaginary field and the real field of a moving object to complete the performance.
However, this does not apply to the alignment of cover and moving objects.
Varcheie et al. [3] proposed a region based BGS method on color histograms and texture information with the GMM to model the background and detection motion. This method improved performance than classical BGS methods, but the intricacy was extremely high. Varadara-jan et al. [4] developed a general structure of GMM by region that considered the pixel dependencies. The framelevel method models the entire image as a vector, rather than simulating changes in individual pixels or regions. The most popular types of approaches are electronic space decomposition and PCA [5 - 8]. In particular, a background model (matrix) is created on behalf of input frames with vector representations in an image frame, so that the correlation matrix is split into eigenvalues. Deep neural networks have become quite popular in recent years. CNN [9 - 12] examined the detailed structure of a large number of benchmark datasets, which significantly improved the performance of the project. However, realtime performance on embedded systems is not possible without a dedicated GPU or processor.
In the process of background subtraction, due to the internal time-domain coupling of the background image, the remodeled matrix has a low order. And the L-1 norm was used to characterize the scarcity in the foreground. But the model works for background subtraction tasks, assuming approximate dynamic and static components analogous to the foreground and background of the video, respectively. Therefore, it is necessary to take into ac-
count the best knowledge to solve emerging problems. Aiming on the restrictions of RPCA-based techniques, this paper presents rank-1 standardized RPCA and TV (Total variation) model that makes full use of the static and dynamic components in front of the information.
This article employs Robust Principal Component Analysis (RPCA) for background subtraction with various methods; Fast PCP, LADMAP for Low-Rank Representation, RPCA with stochastic optimization and RPCA using Inexact
Augmented Lagrange Multiplier. Histogram equalization is used for pre-processing of input image. Feature extraction is done by using cartoon descriptors. Furthermore, median filter with active contour is used for post-processing.
The outline for remaining sections of the paper is structured as follows. The 2nd section offers the methodology for the above technique. Section 3 describes the simulation results for the implemented methodology and Section 4 contains the conclusive remarks with future scope.
Input Histogram TV-
frame equalization L1
Postprocessing
Cartoon Features t —r
i Sparse Matrix
T
Texture RPCA
Feature
\
LowRank Matrix
Fig. 1. Block diagram for proposed hybrid method Proposed methodology
The proposed hybrid method computes background models with different conditions and videos to observe the performance of different RPCA approaches and the effect of preprocessing on results. In this article, we present a fast and computationally efficient algorithm for solving PCP problems with a focus on foreground / background video detection modeling, where the D column resembles for each video frame. Fig. 1 shows the video frame histogram equalization as an initial process.
After that histogram equalization is used to enhance the frame image and TV-L1 decomposition [16] is performed. This process gives two features cartoon features and texture features. This calculated texture features is
sampled for RPCA where the low rank and sparse matrices are computed. The low-ranked matrix (background image) was then resolved using the "lansvd" procedure from the PROPACK optimized library [13]. The sparse matrix was calculated from element-wise compression (soft edge) [14]. Further Sparse matrix is combined with cartoon features and morphological post processing is happened to get moving object from foreground images.
The foreground is achieved by applying postprocessing to the sparse matrix. Fig. 2 shows the postprocessing used. This measurement includes a hard threshold, a morphological filter, and an active border with a negative compression offset [15]. Rest of methodology is explained in following subheadings.
Sparse Matrix
RPCA
Low-rank matrix
Hard Threshold Median Filter
■
Active Contours Opening
1 r
Closing
Opening
B ackground
Foreground
Fig. 2. Block diagram for post-processing [15] 1. Feature extraction using cartoon descriptors
The cartoon feature in images is classically obtained by using a pair of high pass and low-pass filters to the image I by the following minimization [16]:
in|ct4 J| Du |2 +1| I -u |!H-1}. (1)
min
Where u represents the cartoon part of the image I, D represents the first derivative of u, ct2 is variance of image I [16].
The texture and cartoon image parts are achieved by the following equations (referred from Buades et al., [17]):
u (x) = ro(Xa (x))(Ln * I(x)) + (1 - ro(Xa (x)))I(x),
(2)
v(x) = I -u(x).
(3)
Where ra(.) symbolizes the weight function, L a is a low pass filter and Xcis high pass filter [17].
Equation (4) defines the data matrix M obtained by RPCA, ft (x,y) is the function for cartoon texture descriptor extracted from each image. L and S matrices are derived from matrix M by RPCA [18].
M (x, y) = p f (x, y).
2. Fast principal component pursuit
(4)
The classification performance can be improved with the utilization of the fast principal component pursuit method focusing on the incremental PCP algorithm [14] [19, 20].
Various approaches based on histogram [21], subspace training [22] and neural networks [23] are used to implement the background video modeling. The latter model is based, in particular, on PCP [5] [24].
The PCP algorithm is defined by the equation (5) [24]:
arg minl¿rank(L) + X || S ||0 s.t D=L+S.
(5)
Where De Rmxn is the observation video of n images, each with size m = Nr *Nc*Nd (row, column and depth or channel), Le M.mxn is a lower-order matrix demonstrating the background of the plane, and Se Mmx" is a sparse matrix signifying the foreground (moving object).
Although most PCP algorithms directly rely on convex relaxation shown in equation (6):
arg minl,s ||L||* +X || S ||i s.t D=L+S.
(6)
As presented in [25], equation (7) is also a suitable convex relaxation of the equation (1).
arg minLS 1||L+ S-D ||F s.t || S|1 <x,rank(L) <r. (7)
What can be done repeatedly using alternative optimizations:
Lk+v> = arg minl 21| Lk+S®-Dk ||F s.t rank(Lk) <r, (8)
SK+» = arg mins ^2||L(/+1) + Sk - Dk ||F s.t ||Sk ||i <x. (9)
Where Lk = [Lk -1 lk], Sk = [Sk -1 sk] and Dk = [Dk - i dk]. The minimization of equation (6) can be calculated using the incremental thin SVD method [23],while the equation (7) is the prediction (dk - lk) against the sphere li.
The PCP algorithm [14] [18] [19], utilized in this research work, uses a special structure of solutions suggested in [20] to convert them into incremental solutions: solutions (and packages) of computationally intensive problems with partial equations (8) can be computed efficiently by changing the 1st classification for thin SVD [26], which is computed with the availability of a new video
(image).It provides a fully incremental algorithm which is independent of background variations.
Algorithm to Solve Fast PCP Method [26]:
1. Input video D, initialize parameter: S1 = 0, rank =1 (the initial rank)
2. While not converge do
3. Solve Lk+1 with rank =t and preserve singular
values to v
4. Calculate I vmnk / X vk
5. If Ivrank / X vk J >x, then + +rank
6. Solve Sk+i
7. end while
8. Output: L, S.
Since the video background modeling application L is low rank matrix and upper bound of x is controlled to get new singular values Lk+1 from equation 5. If observed contribution are small enough then increment in x is stopped. The proposed simulation x is increased up to 3.
3. Online robust PCA via stochastic optimization
We now present robust PCA using stochastic optimization [27]. The main objective of this approach is to present a cost function minimization method utilizing a stochastic optimization [27] :
1
fn(L) = -Xl(zi,L) +—1| L
" 2n
(10)
Factor r, noise e and L base are optimized as alternatives. At the time t, a base estimate for L, is used by minimizing the accumulated loss with respect to preliminary predictable coefficients Factor r, noise e and L base are optimized as alternatives. At the time t, a base estimate for Lt is used by minimizing the accumulated loss with
respect to preliminary predictable coefficients {ri ^ and
spurious noise {e; }t . The L, database is updated by utilizing the following objective function and spurious noise. The Lt database is updated by utilizing the following objective function [27]:
g (L) = 1 ¿(11| zi - Lr - ei ||2 +|||,||2 +X21| ei ||i j +
+|||L||F.
(11)
It is a substitution function for the empirical cost function ft (L) demarcated in equation (5). This gives an upper bound for ft (L):gt (L) > ft (L).
4. RPCA using inexact augmented lagrange multiplier
In this section we contemplate the nonlinear scalar problem with constraints in the form of:
Minimize f (x ), Subject to x eQ.
(12)
Where / = Mn ^ M is continuously differentiable inRn.Given x*e Q, A(x*) ={je{1,..,p}:g,(x*) = 0} denote to the set of dynamic discrimination constraints on x*.
To solve equation (12), the classical Augmented La-grangian method is easily solved by a succession of subproblems. In each sub-problem, given the fixed penalty parameter p > 0 and estimates of the Lagrange multipliers XeMm, |eMp, | > 0. Although there are several penalty functions [28], the external penalty function used in this work is the quadratic function. Due to its simplicity, the quadratic penalty function is the most used in practice, although from time to time the use of other penalty functions may have advantages, see for example [28] [29] and its references.
Given xeMn, p > 0, XeMm, |eMp, | > 0 the generalized formula of the Augmented Lagrangian function associated with the scalar problem (12) considering the quadratic penalty function is:
L( x, X, p) = f (x) +
+P
a * (x)+£
max <¡0, gj (x) +—
2 V
(13)
Simulation results
In this paper, we compare four background subtraction techniques such as RPCA STOC, LRR-LADMAP, RPCA-IALM and RPCA Fast PCP using the CD Net dataset [30]. The database contains 1700 frames with a resolution of 320x240, the first 100 frames are used for background initialization, and the rest of the images are used for background updates for object detection. Each frame has a separate ground truth. Four techniques of background subtraction are compared. Precision, sensitivity, false positive rate, specificity and accuracy are the parameters used for evaluation of the proposed techniques.
The 2014 DATASET offers a variety of realistic camera images (no CGI) and various videos [31]. They have been selected to address a wide range of detection and presentation challenges for typical internal and external visual data captured in today's video databases, intelligent environments, and surveillance scenarios. The dataset is accompanied by precise ground-truth segmentation and displacement / motion zone annotations for each video frame. Similarity measurements were performed using Euclidean distance.
Fig. 3, 4 and 5 demonstrate the RPCA operation on Escalator and Highway video respectively. The video sequence is decomposed into sparse and low-rank matrices. Further low rank matrices are processed with morphological operation to get the filtered outcome as a moving object detection.
In the background subtraction process, the difference between previous frame and current frame is computed by a parameter which is known as running average. In other way it is the absolute difference between two frames where previous background model and current background model with new object introduced in the background.
(d) * « (e)
Fig. 3. Simulation performed on escalator image a) Input, b) Low rank, c) Sparse, d) Outliers, e) Filtered Outliers
Input (I)
Low Rank (L)
(e)
Fig. 4. Simulation performed on Highway image-1 a) Input, b) Low rank, c) Sparse, d) Outliers, e) Filtered Outliers
rrf;^""" (e) Fig. 5. Simulation performed on Highway image-2 a) Input, b) Low rank, c) Sparse, d) Outliers, e) Filtered Outliers
The running average is calculated using the following equation [35]:
|t = ax, + (1 -a) |t -1.
(14)
Where | is the mean computed up to frame t, a is the learning rate of the model, and xt is the pixel value in frame t.
It can be seen in fig. 6 running average is high at initial video index, as data index points increase the RPCA-PCP performs with lower average error values. Subsequently it can be seen in fig. 7. RPCA-PCP yields very lower residual error.
Tab. 1 represents the comparative analysis of various proposed methods on meet video, with cartoon features.it can be seen Proposed Hybrid Fast PCP methods gives higher accuracy 97.5 % as compare to RPCA+STOC (95 %), RPCA+IALM (94.2 %) and LLR+LADMAP
(96 %). F-Score, MCC and kappa statistics are 97.6 %, 95.1 % and 95.1 % respectively from method RPCA+FASTPCP, which is higher than other proposed methods. Higher MCC, Kappa Statistics and F-Score indicate better classification. Hence Fast PCP along with cartoon feature performs better than other algorithm used in paper.
Tab. 2 represents the comparative analysis of various proposed methods on meet video, with LBP features.it can be seen Proposed Hybrid Fast PCP methods gives higher accuracy 95.01 % as compare to RPCA+STOC (92.5 %), RPCA+IALM (87.5 %) and LLR+LADMAP (94 %).F-Score, MCC and Kappa statistics are 95.2 %, 90.5 % and 90 % respectively from method RPCA+FAST PCP, which is higher than other proposed methods.
From fig. 8, it can be seen clearly that all algorithm incorporated with cartoon feature gives better accuracy than LBP features.
The above tab. 3 represents the comparative analysis of background subtraction in meet video [32] has used wavelet transform based background segmentation which increase the more complex structure and volatile to change detection. This method claim sensitivity to 83.91 %. Authors of [33] used the deep learning methods for background subtraction which may not be applica-tionspecific because of trained network. In general deep learning methods can work in generalized way and may have accuracy compromise, whereas proposed hybrid method with cartoon features yields 98.9 % of sensitivity.
-RPCASTOC RPCA IALM ■LLR FastLADMAP ■Proposed-method
100 200 300 400 data stream index
Fig. 6. Running average error graph
1
0.8 0.6 0.4 0.2 0
1 X
Ik ¡Hi , 11 1 'Vil U| i \
-9-RPCASTOC -RPCA IALM - - LLR FastLADMAP .........Proposed-method
0
400
100 200 300 data stream index
Fig. 7. Normalized residual error graph
Tab. 1. Proposed hybrid method evaluation for Dataset14 (Meet) with cartoon features
RPCA+STOC+Cartoon RPCA+IALM+Cartoon LLR+Fast-LADMAP+Cartoon Proposed Hybrid RPCA+FAST PCP+Cartoon
Accuracy 9.50e - 01 9.42e - 01 0.96 9.75e - 01
Error Rate 5.00e - 01 5.83e - 02 0.22 2.50e - 02
Sensitivity 1.00e + 00 1.00e + 00 1 1.00e + 00
Specificity 0.9 0.8833 0.92 9.50e - 01
Precision 0.909 0.8955 0.926371 9.52e - 01
FPR 0.0999 0.1167 0.07996 5.00e - 02
F-Score 9.52e - 01 9.45e - 01 0.96164 9.76e - 01
MCC 9.05e - 01 8.89e - 01 0.923183 9.51e - 01
Kappa Statistics 9.00e - 01 8.83e - 01 0.91998 9.50e - 01
Tab. 2. Proposed method evaluation for Dataset14 (Meet) with LBP features
RPCA STOC+LBP RPCA IALM+LBP LLR+Fast LADMAP+LBP Proposed Hybrid RPCA FAST PCP+LBP
Accuracy 9.25e - 01 8.75e - 01 0.94 9.50e - 01
Error Rate 7.50e - 02 1.25e - 01 0.06 5.00e - 01
Sensitivity 1.00e + 00 1.00e + 00 1 1.00e + 00
Specificity 8.50e - 01 7.50e - 01 0.88 0.9
Precision 8.70e - 01 8.00e - 01 0.8967 0.909
FPR 1.50e - 01 2.50e - 01 0.12 0.0999
F-Score 9.30e - 01 8.89e - 01 0.94454 9.52e - 01
MCC 8.60e - 01 7.75e - 01 0.88824 9.05e - 01
Kappa Statistics 8.50e - 01 7.50e - 01 0.88 9.00e - 01
Tab. 3. Comparison with previous research works for Meet input video
Methods Sensitivity Precision FPR Specificity
Khare et al. [32] 83.19 % 79.15 % 17.33 % 82.67 %
Dou et al. [33] 93.58 % 93.15 % 3.11 % 96.89 %
Sengar and Mukhopadhyay [34] 91.99 % 95.98 % 07.49 % 92.51 %
Proposed Hybrid Method 98.9 % 95.2 % 5 % 95 %
Accuracy in Percentage
98 96 9+ 92 90 ss 86 84
1
RPCA SI OC KP< JAIALM LLRFAST RPCAFASTPCP ■ LEP CARTOON
Fig.8. Comparative analysis of average Conclusion
Multi-Layer RPCA has been proposed for video background subtraction, using texture descriptors. The proposed algorithm consists of preprocessing, RPCA and post-processing. The preprocessing utilizes Gaussian filtering, histogram equalization, or a combination of both. Cartoon features are used for the texture information, and the background subtraction is accomplished by the low-rank and sparse matrix computation achieved by RPCA algorithm. Post-processing calculates active contours and morphological filters with negative shrinkage offset. The proposed hybrid Multi-Layer RPCA FAST PCP+ Cartoon algorithm produce better results than other cited method in a database. Numerical comparisons suggest that RPCA+ Fast PCP+ Cartoon algorithm has superior performances on synthetic as well as real-world datasets with a maximum accuracy of 97.5 %. It was found that the proposed work attains highest sensitivity (100 %) as compared to the other works. The proposed hybrid algorithm overwhelms the effect of dynamic background and also, it is less complex for real-time processing. In future, the role of deep learning can be explored for background modeling.
References
[1] Jeevith SH, Lakshmikanth S. Detection and tracking of moving object using modified background subtraction and Kalman filter. Int J Electr Comput Eng 2021; 11(1): 217223.
[2] Liu K, Zhang J. Moving object detection based on improved ViBe algorithm. Proc SPIE 2021; 11736: 117360N.
[3] Varcheie PDZ, Sills-Lavoie M, Bilodeau GA. A multiscale region-based motion detection and background subtraction algorithm. Sensors 2010; 10(2): 1041-1061.
[4] Varadarajan S, Miller P, Zhou H. Region-based mixture of Gaussians modelling for foreground detection in dynamic scenes. Pattern Recognit 2015; 48(11): 3488-3503.
[5] Candes EJ, Li X, Ma Y, Wright J. Robust principal component analysis? J ACM 2011; 58(3): 11.
[6] Zhou X, Yang C, Yu W. Moving object detection by detecting contiguous outliers in the low-rank representation. IEEE Trans Pattern Anal Mach Intell 2012; 35(3): 597610.
[7] Gao Z, Cheong LF, Wang YX. Block-sparse RPCA for salient motion detection. IEEE Trans Pattern Anal Mach Intell 2014; 36(10): 1975-1987.
uracy using LBP and Cartoon features
[8] Sun Y, Tao X, Li Y, Lu J. Robust 2D principal component analysis: A structured sparsity regularized approach. IEEE Trans Image Process 2015; 24(8): 2515-2526.
[9] Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierarchies for accurate object detection and semantic segmentation. Proc IEEE Conf on Computer Vision and Pattern Recognition 2014: 580-587.
[10] He K, Zhang X, Ren S, Sun J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 2015; 37(9): 1904-1916.
[11] Ren S, He K, Girshick R, Sun J. Faster R-CNN: Towards real-time object detection with region proposal networks. 2015. arXiv preprint. Source: <https://arxiv.org/abs/1506.01497).
[12] Redmon J, Divvala S, Girshick R, Farhadi A. You only look once: Unified, real-time object detection. Proc IEEE Conf on Computer Vision and Pattern Recognition 2016: 779-788.
[13] Welcome to the PROPACK homepage. Source: <http://sun.stanford.edu/~rmunk/PROPACK).
[14] Rodriguez P, Wohlberg B. Incremental principal component pursuit for video background modeling. J Math Imaging Vis 2016; 55(1): 1-18.
[15] Lo BPL, Velastin SA. Automatic congestion detection system for underground platforms. Proc 2001 Int Symp on Intelligent Multimedia, Video and Speech Processing (ISIMP 2001) (IEEE Cat. No. 01EX489) 2001; 158-161.
[16] Osher S, Solé A, Vese L. Image decomposition and restoration using total variation minimization and the H1. Mul-tiscale Model Simul 2003; 1(3): 349-370.
[17] Buades A, Le TM, Morel JM, Vese LA. Fast cartoon+ texture image filters. IEEE Trans Image Process 2010; 19(8): 1978-1986.
[18] Giraldo-Zuluaga JH, Salazar A, Gomez A, Diaz-Pulido A. Camera-trap images segmentation using multi-layer robust principal component analysis. The Visual Computer 2019; 35(3): 335-347.
[19] Rodriguez P, Wohlberg B. A MATLAB implementation of a fast incremental principal component pursuit algorithm for video background modeling. 2014 IEEE Int Conf on Image Processing (ICIP) 2014: 3414-3416.
[20] Rodriguez P, Wohlberg B. Fast principal component pursuit via alternating minimization. 2013 IEEE Int Conf on Image Processing 2013: 69-73.
[21] Zheng J, Wang Y, Nihan NL, Hallenbeck ME. Extracting roadway background image: Mode-based approach. Transp Res Rec 2006; 1944(1): 82-88.
[22] Oliver NM, Rosario B, Pentland AP. A Bayesian computer vision system for modeling human interactions. IEEE Trans Pattern Anal Mach Intell 2000; 22(8): 831-843.
[23] Maddalena L, Petrosino A. A self-organizing approach to background subtraction for visual surveillance applications. IEEE Trans Image Process 2008; 17(7): 1168-1177.
[24] Wright J, Ganesh A, Rao S, Ma Y. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization. Coordinated Science Laboratory Report no UILU-ENG-09-2210 2009: DC-243.
[25] Rodriguez P, Wohlberg B. An incremental principal component pursuit algorithm via projections onto the I 1 ball. 2017 IEEE XXIV Int Conf on Electronics, Electrical Engineering and Computing (INTERCON) 2017: 1-4.
[26] Brand M. Fast low-rank modifications of the thin singular value decomposition. Linear Algebra Appl 2006; 415(1): 20-30.
[27] Feng J, Xu H, Yan S. Online robust PCA via stochastic optimization. Proc 26th Int Conf on Neural Information Processing Systems (NIPS'13) 2013; 1: 404-412.
[28] Echebest N, Sánchez MD, Schuverdt ML. Convergence results of an augmented Lagrangian method using the exponential penalty function. J Optim Theory Appl 2016; 168(1): 92-108.
[29] Bertsekas DP. Constrained optimization and Lagrange multiplier methods. Academic Press; 2014.
[30] Goyette N, Jodoin PM, Porikli F, Konrad J, Ishwar P. Changedetection.net: A new change detection benchmark dataset. 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops 2012: 1-8.
[31] Wang Y, Jodoin PM, Porikli F, Konrad J, Benezeth Y, Ishwar P. CDnet 2014: An expanded change detection benchmark dataset. Proc IEEE Conf on Computer Vision and Pattern Recognition Workshops 2014: 387-394.
[32] Khare M, Srivastava RK, Khare A. Moving object segmentation in Daubechies complex wavelet domain. Signal Image Video Process 2015; 9(3): 635-650.
[33] Dou J, Qin Q, Tu Z. Background subtraction based on deep convolutional neural networks features. Multimed Tools Appl 2019; 78(11): 14549-14571.
[34] Sengar SS, Mukhopadhyay S. Moving object detection using statistical background subtraction in wavelet compressed domain. Multimed Tools Appl 2020; 79(9): 5919-5940.
[35] Martin C. Background subtraction using running gaussian average: A color channel comparison. Seminar aus Bildverarbeitung und Mustererkennung 2014: 0625040.
Authors' information
Mr. S.H. Jeevith (b.1988), is a Research Scholar at VTU, Belagavi, Karnataka, India. He has perceived B.E (Telecommunication Engineering) and M.Tech (VLSI and Embedded System) from VTU in the year 2009 and 2012 respectively. He is working as Assistant professor, Dept. of Electronics & Telecommunication Engineering, SSIT, Tumakuru, Karnataka, India. He has working experience of more than 8 years. His area of research interest are image processing and VLSI. He is the Associate member of IEI. E-mail: [email protected] .
Dr. S. Lakshmikant. (b.1980) has perceived his B.E (EEE) and M.Tech (CAID) from VTU, Belagavi, Karnataka, India, in the year 2002 and 2007 respectively. He has been awarded Ph.D in Jain University, Bengaluru, Karnataka in the year 2015. He is currently working as Associate professor, Dept. of ECE, Acharya Institute of Technology, Bengaluru, Karnataka. He has working experience of more than 18 years. His area of research interest are signal processing, signal denoising, image processing, power electronics and control systems. He is the member of IEI and ISTE. E-mail: [email protected].
Received September 12, 2021. The final version - May 15, 2022.