iconOpen Access

ARTICLE

crossmark

Visual Object Tracking Based on Modified LeNet-5 and RCCF

Aparna Gullapelly, Barnali Gupta Banik*

Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Deemed to be University, Hyderabad, 500075, Telangana, India

* Corresponding Author: Barnali Gupta Banik. Email: email

Computer Systems Science and Engineering 2023, 46(1), 1127-1139. https://doi.org/10.32604/csse.2023.032904

Abstract

The field of object tracking has recently made significant progress. Particularly, the performance results in both deep learning and correlation filters, based trackers achieved effective tracking performance. Moreover, there are still some difficulties with object tracking for example illumination and deformation (DEF). The precision and accuracy of tracking algorithms suffer from the effects of such occurrences. For this situation, finding a solution is important. This research proposes a new tracking algorithm to handle this problem. The features are extracted by using Modified LeNet-5, and the precision and accuracy are improved by developing the Real-Time Cross-modality Correlation Filtering method (RCCF). In Modified LeNet-5, the visual tracking performance is improved by adjusting the number and size of the convolution kernels in the pooling and convolution layers. The high-level, middle-level, and handcraft features are extracted from the modified LeNet-5 network. The handcraft features are used to determine the specific location of the target because the handcraft features contain more spatial information regarding the visual object. The LeNet features are more suitable for a target appearance change in object tracking. Extensive experiments were conducted by the Object Tracking Benchmarking (OTB) databases like OTB50 and OTB100. The experimental results reveal that the proposed tracker outperforms other state-of-the-art trackers under different problems. The experimental simulation is carried out in python. The overall success rate and precision of the proposed algorithm are 93.8% and 92.5%. The average running frame rate reaches 42 frames per second, which can meet the real-time requirements.

Keywords


1  Introduction

In image processing and computer vision, visual object tracking is an important field of research with several applications in computer vision, human-computer interaction, and CCTV surveillance. The visual object tracking objective is to discover and localize a target of interest in the successive video frames [1]. In this field, advancements have been made through years of scientific research [2]. Due to this visual object tracking, a possibility of achieving multiple difficulties including blur, partial occlusion, camera motion, scale variation, illumination variation, and background clutter. In unconstrained situations, achieving efficient and reliable tracking remains a difficult task [3].

Several tracking algorithms have been developed over the last few years. There are two types of approaches: discriminative and generative approaches [4,5]. As the target location, the tracker determines the best matching region using a generative method, which establishes a target appearance model [6]. The visual tracking is done by the discriminative method in contrast. The classifier is trained using target and background training samples with the vision of making separation among samples [7].

The Discriminative Correlation Filters (DCF) tracking paradigm was created and intensively investigated to reduce the tracking difficulties of standard visual object tracking systems [8,9]. The circulant matrix property ensures the performance of its training and localization phases, which involve entire circularly enhanced samples [10]. Discrete Fourier Transform (DFT) in the frequency domain with closed-form solutions speeds up the correlation operator development, which is represented as a ridge regression problem [11].

Multiple trackers in recent decades have achieved superior performance on benchmarking datasets and in competitions [12]. Furthermore, visual object tracking performance is improved by establishing feature extraction. Feature extraction is the most essential factor for object tracking [13]. When compared to the features of a deep Convolutional Neural Network (DCNN) with descriptors of the conventional images, the features are more effective and intuitive. In the DCF model, several deep features are merged for the object tracking in the video [14,15]. There has not been research into the relationships between several deep channels. In the training framework, the group relationships among deep image features are identified by researching the significance of multi-channel features with high dimensions to achieve adaptive channel selection as a result of this research [16].

In visual tracking, Correlation Filters (CFs) are now commonly utilized due to the significant computing power [17]. Multiple research has been explored that use CFs. Other than generative techniques, tracking algorithms use discriminative techniques such as CFS. The classification problem is considered by discriminative trackers [18]. The relationship between the last frame’s target model and the present frame’s unknown target samples is calculated by using the CFs in the visual object tracking and also the foreground and background are distinguished using CFS [19]. By using single CFs in the visual object tracking, it is ineffective for dealing with fast motion (FM) despite its advantages. Simultaneously, a linear method is used to combine some multiple correlation filter models, which has the drawback of making the tracking algorithms vulnerable to problems like as quick motion and target DEF. To solve these problems, proposed different trackers of deep learning-based but still it has a complex computation limit [20].

In the field of computer vision, visual tracking has attracted a lot of interest since the development of camera technology. A variety of DCF techniques are frequently utilized in tracking, however, the majority of them consistently fail to locate the target in difficult circumstances, failing during the period of the sequences. Some of the existing trackers fail due to motion blur (MB), FM, and occlusions. To handle these issues, propose a different tracking technique that depends on the Modified LeNet-5 feature extraction and RCCF for real-time object tracking for robust object tracking.

This research aims to enhance tracking accuracy while maintaining a reasonable tracking performance under several difficulties. We develop the visual tracking algorithm depending on the Modified LeNet-5 for feature extraction and RCCF. The tracking accuracy is improved by using the RCFF, and increasing the tracking performance by combining a feature fusion factor with a scale estimation factor. As a result, in situations of changes of scale, FM, and DEF, Modified LeNet-5-RCFF can achieve good tracking accuracy.

The novelty of the work is as follows,

•   Features are extracted by using the Modified LeNet-5 model. The target appearance model is represented by feature selection and is based on the middle and high network layers. The different information of the target is expressed by different features, the overall target appearance is represented by combining the different features, which has improved the tracking performance.

•   Then propose a novel RCCF with complex visual changes, unlike previous DCF-based trackers. The function of RCFF is applying correlation filtering in extracted features and generating response maps to enhance visual object tracking.

•   On the OTB50 and OTB100 benchmarks, we perform extensive experiments. The results of the experiment reveal that the proposed tracker performed successfully compared with other state-of-the-art trackers.

The remaining of this research is structured as follows. The related work relevant to the proposed system is discussed in Section 2. The detailed representation of the proposed model and the optimization algorithm is described in Section 3. The experimental results and the performance comparison with various representative and correlative trackers are described in Section 4. Section 5 concludes the paper.

2  Related Works

For real-time object tracking, a new dual-template adaptive correlation filter is proposed by Yan et al. [21]. Different size-level templates are developed first in this paper. Based on the target response confidence, the optimal template was established at second during target translation estimation. The template switching threshold is changed during the object tracking process under each frame’s output response confidence. This can improve the tracker’s ability to adapt to diverse video sequences. Lastly, the Kernelized Correlation Filter (KCF) is developed by integrating the feature fusion factor, scale estimation factor, and dual templates for increasing the tracking performance.

A novel DCF-based tracking system was investigated by Xu et al. [22]. The temporal consistent constraints and adaptive spatial feature selection are two fundamental novelties of the proposed technique, the joint spatial-temporal filter learning is also enabled by this tracker in a lower-dimensional discriminative manifold. More particularly, multi-channel filers were exposed to structured spatial sparsity requirements. As a result, the lasso regularization can approximate the process of learning spatial filters. The global structure is preserved in the manifold by limiting the filter model to a range of values around its historical value and is changed locally for ensuring temporal consistency. Finally, the temporal consistency is selected by proposing a unified optimization framework with preserves the spatial features and developing the discriminative filters with the augmented Lagrangian method.

For visual tracking, the different multiple feature fused model into a correlation filter framework is proposed by Yuan et al. [23] for improving the tracking effectiveness and reliability of the tracker. The model of correlation filter generates different response maps for each feature in different tracking conditions. The noise interference in different features is eliminated by using an adaptive weighting function depending on the corresponding response maps of each feature while maintaining their benefits. Moreover, the fast training and accurate location of the object are provided by using this correlation filter. Furthermore, the scale variation of the target is appropriately handled by developing the effective scale variation detection method in visual object tracking.

Siamese Region Proposal Network (SiamRPN) is proposed by Jain et al. [24] that was described as an offline network with a large dataset. The features are extracted by using multiple sub-networks that are present in this network, such as classification and regression. The visual tracking analysis is enhanced by optimizing the feature extraction model through the combination of the Siamese-RPN++ network and You Only Look Once, Version 3 (Yolo-v3). Previously used recognition frameworks repurposed classifiers or localizers to extract features. Even while object scaling, implements the model to various areas of an image.

An alternate description of visuals was proposed by Zhu et al. [25]. From a target with rapid appearance changes, the dynamic appearance information is extracted by proposing the collaborative representation between successive frames specifically, this has the effect of reducing the background’s undesirable impact. A spatially regularized DCF framework is used to combine the resulting collaborative representation coefficients with the original feature maps for better performance. In this paper, a new feature integration object tracker is proposed by Zhang et al. [26], called correlation filters and online learning (CFOL). The location of the target and their corresponding correlation score is estimated by using the CFOL with the same multi-feature DCF. For online learning, the tracking drifts are reduced by developing the new sampling and updating technique.

A cascade-based tracking algorithm is proposed by Lu et al. [27] for improving the tracker’s robustness and reducing time consumption. Features are extracted by proposing a new deep network first. The feature extraction stage can achieve speeds of up to 50 frames per second due to the use of specific pruning algorithms. Moreover, using more spatial information, the performance of tracking is improved by introducing a Dense Connection-based Cascade Tracker (DCCT). The proposed DCCT tracker, like the cascade classifier, is made up of numerous weaker trackers. The false candidates of the tracked object are rejected by each weak tracker and integrated with these weak trackers for obtaining the final tracking results.

3  Proposed Methodology

The target object is tracked more accurately across the entire frame while avoiding high overflow, this is the main aim of our system. For tracking objects in a video frame, the system uses Modified LeNet-5 and RCCF methodologies with various challenge factors including Low Resolutions (LR), Out of Views (OV), Scale Variations (SV), MB, Out of Plane Rotations (OPR), Occlusions, Illumination Variations (IV), FM, DEF, In-Plane Rotations (IPR), and Background Clutter (BC). The schematic diagram of the proposed method is given in Fig. 1.

images

Figure 1: Schematic diagram of our proposed methodology

The target appearance model is represented by proposing a Modified LeNet-5 for feature extraction and the extracted features are selected based on the middle and high network layers. The different information of the target is expressed by different features and the overall appearance of the target is represented by combining these features for improving the performance of the tracking. For visual tracking with complex appearance changes, we propose a new selective correlation filter model termed RCCF for object tracking in the crowd that contributes to improved tracking. The correlation filter generates a response map corresponding to each feature in the RCCF algorithm. The noise interference in different features is eliminated by using an adaptive weighting function depending on the corresponding response maps of each feature while maintaining their benefits. It effectively improves the tracker’s tracking performance and reliability.

After feature extraction, three different filter kernels are used to map the extracted feature by using the function of RCFF. With the corresponding kernel on three levels of image features, the three correlation maps are generated by performing correlation filtering. Lastly, we employ pixel-wise averaging to combine the three correlation maps. The fused heat map’s peak value corresponds to the central point. Then performing the Size and Offset Regression, only the last-level image feature is used to regress the local offset and two-dimensional object size for the center point. Combining the local offset, estimated center point of the object, and size of the object provides the target object region.

3.1 Feature Extraction using Modified Lenet-5

From the visual data, the object is extracted in the form of frame sequences, which is called Visual tracking. It is a popular method for identifying an object when the appearance of the object changes. And the target object is tracked across multiple frames without knowing anything about the target object. However, the extraction model recognizes the initial frame sequence and based on that, the model is effectively followed from the first to the last frame. An original video is divided into several frames, and the target area has been manually modified in the LeNet-5 network also known as a region of interest or an area of interest. This method depends on the model of cross-correlation which is the local feature extraction technique embedded with appropriate classes. Using LeNet-5, the proportional region is obtained by comparing the appropriate frames. This network generates feature templates, which are then compared to the object feature to improve model tracking.

The object is also tracked with different scores by using different bounding boxes. The object’s specific score is generated by the system which is reciprocated with 2k logits. The object is classified by the trained network using full proportional segmentation and binarization and LeNet has been upgraded with LeNet-5, which eliminates network faults and the features are analyzed by exploring more strategic behavior for improved object detection.

The LeNet-5 was made up of seven layers such as two pooling layers, two fully connected layers, two convolution layers, an input layer, and an output layer. An input image {Ai,bi}i=1N is taken formally, where the original image data is represented by Ai , and the image’s class category (i.e., 0 and 1) is represented by bi . The categorical cross-entropy function is used to calculate the difference between the actual label bi and the predicted label bi , it is given by:

F(s,t)=Δ1Nl=1Nbl1logbl1+....+blklogblk (1)

The biases and weights of the conventional LeNet-5 network layers are represented by t and s, respectively. The range of class category is depicted by K and the softmax value of the k’s class category corresponds to blk , defined as:

blk=softmax(zk)=ezki=1Kezi (2)

where the last fully connected layer’s corresponding ith class category result is defined zi . The loss derivative is back-propagated (BP) for factors over the whole network and was used to learn the convolutional operation’s weight and bias parameters.

In the Modified LeNet-5 network, the fully-connected layer nodes and convolution layer strides are not suitable for feature maps operation. Therefore, we made the following changes to LeNet-5: (1) Rather than a 2-dimensional convolution operation, feature extraction in the network utilizing a one-dimensional convolution operation; (2) The over-fitting problem is solved by an additional dropout layer between the fully connected and the convolution layers; (3) The network complexity is reduced by maintaining only one fully connected layer (4) adjusting the fully-connected layer node’s range and the convolution layer stride’s size. In the modified LeNet-5, the feature maps were increased layer by layer, and all convolution layer strides were changed to two when compared to the standard LeNet-5. Between the fully connected layer and convolution layer, a 0.8 drop rate of the dropout layer was added in particular and the binary classification problem is solved by reducing the nodes in the output layer from ten to two.

3.2 Correlation Filtering using RCCF Method

The discrimination between images and their changes is determined by performing correlation filtering through linear template training. In several computer vision fields, correlation filtering is generally implemented. The classification of objects can be considered as a correlation filtering operation, where the filter kernel is the image feature vector output that performs correlation filtering on the last multi-layer perceptron’s weight matrix. It attempts to locate an object in a video based on its location in the first frame for single object tracking, when comparing the initial frame to the others, correlation filtering is helpful.

The proposed RCCF approach for visual object tracking is described in this section. Our goal is to determine the exact location of the object depicted in the image without the process of generating proposals. We first perform Image feature correlation filtering with an image-guided filter kernel to locate the object center point in the process of RCCF. After that, the size of the object and offset of the center point are then regressed using a regression module. The target bounding box is formed by combining the correlation heat map’s peak value, regressed object size, and center point offset.

3.2.1 Framework

Assume Q is the object of the image and the image height H and image width W are denoted by the IRS×T×3 . The object size (su,tu) and the center point (au,bu) represent the target object region. There are two modules in the proposed RCCF i.e., correlation filtering and regression modules of size and offset. IF and V1 represents the middle-level and high-level features respectively. The M() shows the cross-modality mapping function then it maps the middle-level feature into the visual domain. The Map(IF) mapping result as the filter (kernel) with V1 a visual feature map, is used to convolve the correlation filtering module and generates the heat map CorrUTdxSd , where the output stride is represented by d. The (x,y) represents object point expressed by the target is indicated by the correlation peak value. Furthermore, the center point’s local offset (αa,αb) and object size (s,t) is predicted by the regression module of the size and offset.

The three levels of visual features are used in this process. The three-level features [VI1,VI2,VI3]=V(I) from Modified Lenet-5 are extracted and converted into a uniform size Td×Sd from T8d×S8d×T4d×S4d and T2d×S2d . All 3-level features are used, when computing the correlation map C because the size [VI1,VI2,VI3] is all 64×Td×Sd . For computational efficiency, only the VI1 highest resolution is used during the regression process. The two modules will next be described in detail.

3.2.2 Cross Modality Correlation Filtering

The method is used to locate the target box center (x,y) . The image-guided kernel generation, correlation operation, and correlation map fusion are the major three steps in RCCF. At first, [k1,k2,k3]=[Map1(IF),Map2(IF),Map3(IF)] is generated from the expression feature IF using three different linear functions.

From the visual space of the target, the cross-modality mapping function projects using the three connected layers [Map1(IF),Map2(IF),Map3(IF)] . A feature vector represents each kernel with a vector size of 64 dimensions. After that, it’s reshaped into a 64 × 1 × 1 filter for further process. On the three levels of visual features, the correlation operations are performed by using mapping kernels [C1,C2,C3]=[k1VI1,k2VI2,k3VI3] , where the convolution operation is denoted as . Finally, pixel-wise averages of the three correlation maps are input into an activation function C=Sigmod(C1+C2+C33) . C1, C2, and C3 are all RTd×Sd in size. The target object’s center point is the location with the highest C score. An output stride d is used to calculate the low-resolution equivalent (ag,bg)=[(ag,bg)d] for the center point of the ground truth (ag,bg) . In the heat map C[0,1]Wd×Hd , the ground truth center point is splatted by using the Gaussian kernel Cab=exp((aag)+(bbg)22σt2) . At the spatial location (a,b) , the value of C is denoted by Cab the object size’s standard deviation σt . A regression process is done pixel-wise with focal loss that is penalty-reduced, which is the training objective.

LC=ab{(1Cab)αlog(Cab)ifCab=1(1Cab)β(Cab)log(1Cab)otherwise} (3)

ReLU on a 3-dimensional convolutional layer is used in both the offset and size regression branches. During training, used the Loss1 loss function. The Lossoff (local offset regression loss) and Losssize (object size regression loss) are defined as:

Lossoff=|Δaagbgδag|+|Δbagbgδbg| (4)

Losssize=|Sagbgsg|+|Tagbgtg|

where the target box’s height and width for ground truth are represented by tg and sg , the ground truth offset vector is denoted by δag=(agdag) and δbg=(bgdbg) . While Hagbg,Δxagybg,andΔyagbg are defined similarly. All other locations are ignored by the regression loss, which processes at the center point location (ag,bg) . The weighted sum of three-loss terms provides the final loss, which is given in below equation

Loss=LC+λsizeLosssize+λoffLossoff (5)

Both values of λsize and λoff are set to 1, and a normalized object size coefficient is represented by λsize . The target center point (at,bt) in the heat map C is selected based on the highest confidence score during inference. The target offset and size are calculated using the relevant location in the Δaat,bg,Δbat,bt,ΔaandΔbasSat,bt,Tat,bt,S,T . The bottom right and top left corner coordinates of the target box are calculated by:

at+Δaat,btSat,bt2,bt+Δbat,btTat,bt2,at+Δaat,bt+Sat,bt2,bt+Δbat,btTat,bt2 (6)

4  Experimental Results

We evaluate our tracker using a standard visual tracking benchmark to objectively and properly examine the proposed tracker. We describe the algorithm flow as well as the experimental details and specifics. Then we provide the experimental evaluation’s details and standards. Finally, the OTB50/OTB100 benchmarks, which contain over 100 test video sequences, are used to validate the performance of our tracker. And our methodology is implemented in Python.

4.1 Datasets

Well-known benchmarks such as OTB50 and OTB100 are examined to assess the performance of our tracking approach. OTB50 and OTB100 datasets are commonly utilized. The dataset contains 11 attributes including LR, OV, SV, MB, OPR, Occlusions, IV, FM, DEF, IPR, and BC. There are many sequences in each of them.

4.2 Evaluation Metrics

The tracking algorithms are evaluated by the OTB dataset on two levels such as success plot and precision plot. The average Euclidean distance between the ground truth and target prediction center location is denoted by CLE. In the object tracking in videos, the percentage of the entire frame is used to calculate the Distance Precision (DP), with comparing the set threshold, the CLE was smaller in the tracking process. The DP threshold was set to 20 pixels for this research.

The one-pass evaluation (OPE) protocol function is starting with the first frame’s ground truth position and it provides the performance of success rate and average precision. Based on bounding box overlap and center location error, success and precision charts are provided. The factors employed in the evaluation were the DP, Overlap Precision (OP), and Area under Curve (AUC).

The tracker’s running speed is also an important factor to consider while evaluating its performance. Running speed is measured in frames per second (FPS).

4.3 Implementation Results

In the first frame, the different features are first extracted by the specified initial bounding box, and then corresponding filters are trained using the proposed methodology. In the tracking sequences, the tracker is then iterated over each frame. The suitable scale size and the location of the target’s center position are determined by using the several features fused model in each iteration successively. The correlation filter models are finally updated progressively. The proposed methodology section shows the entire process of the method. The average frame rate for a running frame is 42 FPS.

The proposed tracker outperformed the existing trackers with multiple attributes, i.e., LR, OV, SV, MB, OPR, Occlusions, IV, FM, DEF, IPR, and BC. And also the robustness of the proposed method is validated by determining the attribute-based evaluation. While analyzing the AUC and DP scores on these attributes, the proposed tracker performed better on these factors than all competing trackers. These results demonstrate that the proposed tracker is more robust than the other trackers.

The experimental results of the proposed visual object tracking model are shown in Figs. 2 and 3 for the OTB50 and OTB100 datasets, respectively.

images

Figure 2: The experimental results of the proposed visual object tracking using the OTB50 dataset

images

Figure 3: The experimental results of the proposed visual object tracking using the OTB100 dataset

4.4 Performance Evaluation

Scale Adaptive with Multiple Features tracker (SAMF), KCF, Discriminative Scale Space Tracking (DSST), and Dimension Adaption Correlation Filters (DACF) are the recently proposed algorithms that are compared to our proposed methodology. For the OTB-50 and OTB100 datasets, Fig. 4 depicts the overall performance analysis, and it shows the performance of the success and precision plots. When the threshold is set to 20 pix, the algorithms are ordered by average DP in the precision plot. When the threshold is 0.5, the algorithms are sorted by AUC in the success plot.

images

Figure 4: OPE success plot and precision plot on OTB-50 and OTB-100

For OTB50, the proposed method attains the success rate and precision is 0.72 and 0.89 and for OTB 100, the proposed method achieves the success rate and precision is 0.63 and 0.8, comparing the results of the proposed method with recent existing methods such as DACF, SAMF, KCF, and DSST, our tracker achieves better performance and the visual object is tacked efficiently.

The performance comparison of the proposed method and state-of-the-art tracking methods are shown in Table 1, On the OTB-50 and OTB-100 datasets, we compare the performance of our proposed tracker with state-of-the-art trackers in terms of overlap success rate (OSR) and at an overlap threshold of 0.5 and distance precision rate (DPR) at a threshold of 20 pixels in Table 1, We can observe from this table that the proposed tracker produces effective tracking results. Compared to the proposed tracker with different existing trackers SAMF, KCF, DSST, and DACF, the proposed tracker has attained enhanced performance of tracking than these trackers. These advantages are due to our hybrid Modified LeNet-5 and RCCF tracking methodology.

images

5  Conclusion

In computer vision, object tracking in visual is a popular topic. When the target deforms or moves rapidly, the existing correlation filter-based trackers frequently experience drift of the tracking box or the tracking target loss. Focusing on this problem, we proposed a tracking method based on the Modified LeNet-5 feature extraction and RCCF for real-time object tracking. The problems mentioned above can be solved with our tracking technique. The middle and high-level features of the object represent the target model, which shows it enhances the visual object tracking. Furthermore, the target is tracked by using the RCCF, which has been trained on a specific feature. These correlation filters provide complementary response maps. The final response map is improved by fusing several response maps with the RCCF method, where the final tracking result is the maximum value of the response map. A large-scale benchmark named OTB50 and OTB100 is used to conduct extensive research. The overall precision and success rate for OTB50 is 95.2% and 94.51%, and for OTB100 is 89.3% and 93.1%. The results of the experiments reveal that our proposed tracker outperforms other state-of-the-art tracking algorithms. We would refine our method in the future and investigate if this feature is suitable for other trackers.

Funding Statement: The authors received no specific funding for this study.

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

 1.  J. Yang, H. Ge, J. Yang, Y. Tong and S. Su, “Online multi-object tracking using multi-function integration and tracking simulation training,” Applied Intelligence, vol. 52, no. 2, pp. 1268–1288, 2022. [Google Scholar]

 2.  Y. Yu, L. Chen, H. He, J. Liu, W. Zhang et al., “Second-order spatial-temporal correlation filters for visual tracking,” Mathematics, vol. 10, no. 5, pp. 684, 2022. [Google Scholar]

 3.  X. Cheng, Y. Zheng, J. Zhang and Z. Yang, “Multitask multisource deep correlation filter for remote sensing data fusion,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 13, no. 1, pp. 3723–3734, 2020. [Google Scholar]

 4.  A. M. Roy, R. Bose and J. Bhaduri, “A fast accurate fine-grain object detection model based on YOLOv4 deep neural network,” Neural Computing and Applications, vol. 34, no. 5, pp. 3895–3921, 2022. [Google Scholar]

 5.  T. Yang, C. Cappelle, Y. Ruichek and M. El Bagdouri, “Multi-object tracking with discriminant correlation filter based deep learning tracker,” Integrated Computer-Aided Engineering, vol. 26, no. 3, pp. 273–284, 2019. [Google Scholar]

 6.  Z. Zhong, Z. Yang, W. Feng, W. Wu, Y. Hu et al., “Decision controller for object tracking with deep reinforcement learning,” IEEE Access, vol. 7, no. 1, pp. 28069–28079, 2019. [Google Scholar]

 7.  Q. Liu, D. Chen, Q. Chu, L. Yuan, B. Liu et al., “Online multi-object tracking with unsupervised re-identification learning and occlusion estimation,” Neurocomputing, vol. 483, no. 1, pp. 333–347, 2022. [Google Scholar]

 8.  F. Bastani, S. He and S. Madden, “Self-supervised multi-object tracking with cross-input consistency,” Advances in Neural Information Processing Systems, vol. 34, no. 1, pp. 229–254, 2021. [Google Scholar]

 9.  J. Zhang, J. Sun, J. Wang and X. G. Yue, “Visual object tracking based on residual network and cascaded correlation filters,” Journal of Ambient Intelligence and Humanized Computing, vol. 12, no. 8, pp. 8427–8440, 2021. [Google Scholar]

10. D. Elayaperumal and Y. H. Joo, “Aberrance suppressed spatio-temporal correlation filters for visual object tracking,” Pattern Recognition, vol. 115, no. 1, pp. 107922, 2021. [Google Scholar]

11. D. Yuan, X. Shu and Z. He, “TRBACF: Learning temporal regularized correlation filters for high performance online visual object tracking,” Journal of Visual Communication and Image Representation, vol. 72, no. 1, pp. 102882, 2020. [Google Scholar]

12. J. Fan, H. Song, K. Zhang, K. Yang and Q. Liu, “Feature alignment and aggregation siamese networks for fast visual tracking,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 14, pp. 1296–1307, 2020. [Google Scholar]

13. L. Zhou, J. Li, B. Lei, W. Li and J. Leng, “Correlation filter tracker with sample-reliability awareness and self-guided update,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 5, no. 1, pp. 23–56, 2022. [Google Scholar]

14. Y. Huang, Z. Zhao, B. Wu, Z. Mei, Z. Cui et al., “Visual object tracking with discriminative correlation filtering and hybrid color feature,” Multimedia Tools and Applications, vol. 78, no. 24, pp. 34725–34744, 2019. [Google Scholar]

15. J. Shin, H. Kim, D. Kim and J. Paik, “Fast and robust object tracking using tracking failure detection in kernelized correlation filter,” Applied Sciences, vol. 10, no. 2, pp. 713, 2020. [Google Scholar]

16. J. Zhang, X. Jin, J. Sun, J. Wang and A. K. Sangaiah, “Spatial and semantic convolutional features for robust visual object tracking,” Multimedia Tools and Applications,” vol. 79, no. 21, pp. 15095–15115, 2020. [Google Scholar]

17. A. Bathija and G. Sharma, “Visual object detection and tracking using Yolo and sort,” International Journal of Engineering Research Technology, vol. 8, no. 11, pp. 345–355, 2019. [Google Scholar]

18. P. Nousi, D. Triantafyllidou, A. Tefas and I. Pitas, “Re-identification framework for long term visual object tracking based on object detection and classification,” Signal Processing: Image Communication, vol. 88, no. 1, pp. 115969, 2020. [Google Scholar]

19. S. Kanagamalliga and S. Vasuki, “Contour-based object tracking in video scenes through optical flow and gabor feature,” Optik, vol. 157, no. 1, pp. 787–797, 2018. [Google Scholar]

20. M. Yang, Y. Wu and Y. Jia, “A hybrid data association framework for robust online multi-object tracking,” IEEE Transactions on Image Processing, vol. 26, no. 12, pp. 5667–5679, 2017. [Google Scholar]

21. J. Yan, L. Zhong, Y. Yao, X. Xu and C. Du, “Dual-template adaptive correlation filter for real-time object tracking,” Multimedia Tools and Applications, vol. 80, no. 2, pp. 2355–2376, 2021. [Google Scholar]

22. T. Xu, Z. H. Feng, X. J. Wu and J. Kittler, “Learning adaptive discriminative correlation filters via temporal consistency preserving spatial feature selection for robust visual object tracking,” IEEE Transactions on Image Processing, vol. 28, no. 11, pp. 5596–5609, 2019. [Google Scholar]

23. D. Yuan, X. Zhang, J. Liu and D. Li, “A multiple feature fused model for visual object tracking via correlation filters,” Multimedia Tools and Applications, vol. 78, no. 19, pp. 27271–27290, 2019. [Google Scholar]

24. I. Jain and S. K. Sharma, “Convolutional siamese-RPN++ and Yolo-v3 based visual tracking regression,” Journal of Scientific Research, vol. 66, no. 1, pp. 1–9, 2022. [Google Scholar]

25. X. F. Zhu, X. J. Wu, T. Xu, Z. H. Feng and J. Kittler, “Complementary discriminative correlation filters based on collaborative representation for visual object tracking,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 2, pp. 557–568, 2020. [Google Scholar]

26. X. Zhang, G. S. Xia, Q. Lu, W. Shen and L. Zhang, “Visual object tracking by correlation filters and online learning,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 140, no. 1, pp. 77–89, 2018. [Google Scholar]

27. Y. Lu, Y. Yuan and Q. Wang, “A dense connection based network for real-time object tracking,” Neurocomputing, vol. 410, no. 1, pp. 229–236, 2020. [Google Scholar]


Cite This Article

APA Style
Gullapelly, A., Banik, B.G. (2023). Visual object tracking based on modified lenet-5 and RCCF. Computer Systems Science and Engineering, 46(1), 1127-1139. https://doi.org/10.32604/csse.2023.032904
Vancouver Style
Gullapelly A, Banik BG. Visual object tracking based on modified lenet-5 and RCCF. Comput Syst Sci Eng. 2023;46(1):1127-1139 https://doi.org/10.32604/csse.2023.032904
IEEE Style
A. Gullapelly and B.G. Banik, “Visual Object Tracking Based on Modified LeNet-5 and RCCF,” Comput. Syst. Sci. Eng., vol. 46, no. 1, pp. 1127-1139, 2023. https://doi.org/10.32604/csse.2023.032904


cc Copyright © 2023 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 893

    View

  • 568

    Download

  • 0

    Like

Share Link