Startseite Video face target detection and tracking algorithm based on nonlinear sequence Monte Carlo filtering technique
Artikel Open Access

Video face target detection and tracking algorithm based on nonlinear sequence Monte Carlo filtering technique

  • Yunming Du , Yi Liu und Jing Tian EMAIL logo
Veröffentlicht/Copyright: 23. Oktober 2023
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

In order to achieve facial object detection and tracking in video, a method based on nonlinear sequence Monte Carlo filtering technology is proposed. The algorithm is simple, effective, and easy to operate, which can solve the problems of scale change and occlusion in the process of online learning tracking, so as to ensure the smooth implementation of learning effect evaluation. Experimental methods should be added to the article summary section. The results show that the algorithm in this study outperforms the basic KCF in terms of evaluation accuracy and success rate, as well as outperforms other tracker algorithms in benchmark, achieving scores of 0.837 and 0.705, respectively. In terms of overlapping accuracy, the reason why this study’s algorithm is higher than KCF is that this study determines the tracking status of the current target by calculating the primary side regulated (PSR) value when the target is obscured or lost, which does not make the tracking error to accumulate. The tracking algorithm in this study is not ranked first in the two attributes of motion blur and low resolution, but the rankings of all other nine attributes belong to the first. Compared with the KCF algorithm, the accuracy plots for the three attributes of scale change, occlusion, and leaving the field of view are improved by 10.26, 13.48, and 13.04%, respectively. Thus, it is proved that the method based on nonlinear sequence Monte Carlo filtering technology can achieve video facial object detection and tracking.

1 Introduction

The tracking and recognition of video faces is a core problem in the field of computer vision, and the technology has attracted more and more attention from researchers in recent years, mainly because the tracking and recognition of video faces has a wide range of application prospects, such as video conferencing, human–computer interaction, forensic identification, video surveillance, and access control. The research purpose of video face tracking and recognition is to simulate the human visual motion perception function, to give the machine the ability to recognize the moving face in the sequence image, and to provide important data basis for video analysis and understanding. The tracking and recognition of video face is often very difficult because of the various changes in the background and face. Although people have conducted extensive research on video face tracking and recognition, and put forward some effective tracking and recognition methods, it is still difficult to develop a set of robust tracking and recognition algorithms for various face changes in video sequences, such as illumination, expression, posture changes, and partial occlusion. Video face tracking and recognition system mainly includes two aspects of face tracking and recognition, tracking and recognition are complementary processes. This study focuses on how to effectively improve the robustness and accuracy of video tracking and recognition. Under the condition that the training data are limited static image and the test data are video sequence, the study focuses on how to use the same model for tracking and recognition, update the surface model at the same time, and use particle filter to organically combine the two into a whole to complete tracking and recognition at the same time. When both the training data and the test data are video sequences, the study focuses on how to effectively improve the recognition accuracy of faces under the condition of changes in expression, posture, and illumination through robust feature extraction method. Meanwhile, the same surface model is used in the tracking and recognition process to improve the tracking and recognition performance.

Another important part of video tracking algorithm is motion estimation algorithm, which mainly includes Kalman filter algorithm, particle filter algorithm, and various improved algorithms on this basis. Kalman filter is a kind of predictor commonly used in signal processing field. It is a linear recursive filter and an optimal linear estimator based on the minimum mean square error criterion. Kalman filter estimates the current value according to the previous time estimate and the current time observation data, using the state equation and recursive method to estimate the current value. The calculation of Kalman filter is small, it can meet the requirement of real-time, and the estimated result is optimal. However, Kalman filter also has great limitations, it requires that the equation of state must be linear and the noise must be Gaussian. However, in practical applications, strictly speaking, all systems are nonlinear, so the study of nonlinear filtering is very meaningful. One of the classical algorithms in the field of nonlinear filtering is extended Kalman filter (EKF), whose basic idea is to perform first-order Taylor expansion on the nonlinear model around the state estimate, and then apply Kalman filtering.

Face tracking is one of the hotspots in computer vision research, and it has broad application prospects in many fields such as security checks, visual surveillance, video conferencing, human–computer interaction, and so on. Visual tracking can usually be seen as solving the problem of estimating the hidden state variables of a system under a given observation. Bayesian estimation method combines the prior knowledge of unknown (implicit) state quantity and the likelihood function describing the relationship between observation and state quantity, and uses Bayesian formula to obtain the posterior probability of unknown quantity. However, iterative Bayesian estimation only has analytical solutions under specific models and assumptions, including linear Gaussian state space models (Kalman filtering) and finite state space hidden Markov models (hidden Markov filtering). However, in many practical problems, the state space model contains nonlinear and non-Gaussian components, so there is no closed optimal solution.

With the development of information technology and the popularity of Internet technology, the generation, transmission, and storage of video and image data have become more convenient. Massive amounts of video and image data are generated every day. How to efficiently process and utilize these video and image data are problems that need to be solved to realize the automation of information processing. Computer vision technology, which can process video and image data and extract useful information from them, is gradually changing people’s lifestyles and playing an increasingly significant role. Thanks to the improvement of computer computing power and the emergence of large-scale datasets, computer vision technology has become the hottest research direction in the field of artificial intelligence at present. Visual target tracking, as an important subproblem of computer vision, has been of great interest to both academia and industry. The visual target tracking task is to process video sequences to automatically target and obtain information about their position and state in the video sequence. Therefore, visual target tracking techniques play a key role in application scenarios involving continuous acquisition of target states [1,2].

The ubiquitous video surveillance system is one of the infrastructures for realizing smart cities and building a safe society, which reduces the occurrence of crime incidents and ensures people’s safety to a certain extent. Countless surveillance cameras generate huge amount of surveillance video data every moment. When an unexpected event occurs, visual target tracking technology can help lock and monitor the relevant target, and further determine the target identity information and analyze and understand the target behavior through other computer vision technologies. In highway traffic monitoring, visual target tracking technology can track the target vehicle and combine with other computer vision technologies to achieve full coverage of the road network for vehicle detection identification and tracking. At the same time, visual target tracking technology can also perform vehicle statistics, traffic flow monitoring, and traffic accident detection. In new retail application scenarios, visual target tracking technology can be used to count the number of customers, depict customer movement trajectories, and track specified customers. In human–computer interaction systems, accurate gesture tracking, limb tracking, and face tracking are the basis for achieving subsequent functions. In defense military applications, target tracking technology is the key technology for end-to-end guidance, and robust and fast tracking algorithms determine the success or failure of guidance [3,4].

In the past three decades, many suboptimal schemes have been proposed, such as EKF, deterministic numerical integration method, etc. But as the dimension of the state increases, the convergence rate of the approximation error decreases. Since the 1990s, Monte Carlo methods have been widely applied under more general and complex conditions. Unlike EKF, sequential Monte Carlo sampling is not an approximation model to make it conform to a given type of posterior probability, it uses weighted sampling values to approximate the true posterior probability value. It has different titles in different application fields, such as particle filter, condensation, etc.

2 Review of literature

In recent years, with the emergence and development of signal processing theory and computer, people have begun to use cameras to obtain environmental images and convert them into digital signals, and use computers to process visual information, which forms the subject of computer vision. Computer vision is a very active field in computer science today. The basic assumption of the discipline of computer vision is that human visual mechanisms can be simulated computationally. The ultimate goal of computer vision research is to enable computers to observe and understand the world through vision like humans, and have the ability to adapt to the environment independently. It is a highly interdisciplinary subject, involving computer science, psychology, physics, signal processing, and applied mathematics. In the past two decades, with the development of various disciplines and computer technology, computer vision has made vigorous development and has been widely used in various fields. Automatic face recognition is an important biometric recognition technology. Compared with other identification methods, face recognition is direct, friendly, and convenient, so the study of automatic face recognition is not only of great application value but it also has important theoretical significance. In recent years, more and more researchers have paid attention to the tracking and recognition technology of video faces. This is mainly due to two reasons: On the one hand, the rapid decline of computer computing and storage costs makes it possible to capture and store image sequences at or near video rates, and on the other hand, the extremely broad market application prospect of video tracking and recognition technology is also the main driving force to promote this research. Video tracking and identification technology in addition to the intelligent video surveillance system has a very important application in video conferencing, human–computer interaction, access control, home entertainment and information security, and many others.

Summarizing the various algorithms that have emerged over the decades, visual target tracking algorithms can be divided into two main categories, namely, motion modeling-based target tracking and appearance modeling-based target tracking.

The research on facial tracking in China began in the past decade and has developed rapidly. There are also some research results that have been or are being converted into commercial software, but there are no relevant reports on the use of such software. According to recent research results in China, Ai Haizhou and the State Key Laboratory of Intelligent Technology and Systems of Tsinghua University conducted a test on the actual environment, and used the differential image method to locate and track the portraits in front of the office desk. They ran for 5 min each time, totaling 1,123 frames 128 × 128 images or 2,610 frames 64 × 64 images, with a detection accuracy of over 97%. The Institute of Automation of the Chinese Academy of Sciences has realized face tracking in the face deformation environment by using the affine transformation method. In addition, some universities, such as Beijing University of Technology, Harbin Institute of Technology, Shanghai Jiao Tong University, and Northwest University of Technology, are also committed to the research and application development of facial tracking technology. At the same time, a large number of articles in this area have been published at important international conferences such as FG, RATFG-RTS, International Conference on Image Processing, and Conference on Computer Vision and Pattern Recognition. However, there are not many applications that can truly be implemented, and generally there is no systematic provision for the overall implementation in the application environment.

Motion patterns are an important feature of targets in time-series images. Motion modeling is based on certain a priori knowledge to analyze the historical motion patterns of the target and predict the next motion patterns of the target. In the early stage of computer vision technology development, target tracking based on motion modeling occupied a mainstream position due to low computational performance and simple application scenarios. As the computational performance improved, the demand for the application of this technology for target tracking was no longer limited to simple motion patterns. Complex motion patterns are often difficult to be accurately predicted, and appearance modeling-based target tracking ignores the estimation of target motion patterns and gradually becomes mainstream [5].

Appearance modeling-based target tracking consists of two parts: target feature modeling and statistical learning modeling. Target feature modeling is the extraction of visual features of the target and the quantitative representation of the target. The visual features can be the most straightforward grayscale, color, and histogram features. They can be hand-designed local binary pattern (LBP), histogram of oriented gradient (HOG), and scale-invariant feature transform features, and features learned by statistical learning, such as convolutional neural network (CNN). Statistical learning modeling is based on the target feature model, and the target is modeled by statistical learning methods. The common statistical learning methods used in visual target tracking tasks are boosting, support vector machine (SVM), correlation filtering, and deep learning. In the visual target tracking task, the purpose of statistical learning modeling is to measure the similarity between the target image block and the candidate image block. Complications such as occlusion, deformation, and illumination changes may occur during the tracking process, and these can affect the feature model of the target and impact the performance of the tracking algorithm. An effective online learning strategy can alleviate the problem of tracking algorithm failure caused by changes in the target feature model during tracking. However, there may be cases of target out of field of view and drastic deformation during the tracking process. In this case, an improper online learning strategy can make the model update negatively and lead to the problem of model drift. A complex appearance model is more beneficial to target tracking than a simple appearance model, but it also brings a greater computational burden and affects the speed of the tracking algorithm. Offline learning saves the processing time of the tracking algorithm compared to online learning, but also ignores the changes in the target feature model during the tracking process, sacrificing the tracking accuracy. Therefore, building more robust and efficient target feature models and statistical learning models is a research focus and research trend in the field of visual target tracking [6,7]. The face is a type of natural structural object with quite complex detailed changes. The challenge of object detection and tracking lies in the variability of patterns in the face itself due to differences in appearance, expression, skin color, etc. Generally speaking, the image of a face that may have accessories such as glasses and whiskers as three-dimensional objects is inevitably affected by shadows generated by lighting. Therefore, if we can find solutions to these problems and successfully construct a face detection and tracking system, it will provide important insights for solving other similar complex pattern detection and tracking.

Face detection and tracking have become key technologies in computer vision and related fields, with broad application prospects and commercial value in intelligent human–machine interaction, security monitoring, video conferencing, medical diagnosis, and content-based image storage and retrieval.

In this study, a face tracking algorithm based on SMCF technique is presented. The face tracking algorithm suitable for evaluating the learning effect is given based on the previous research and validated on the OTB dataset with very good results. Monte Carlo filtering face tracking algorithm based on sampling theory is proposed. By measuring the similarity between expected samples and target samples, sample weights are calculated to achieve posterior prediction and estimation. The algorithm enhances the robustness and adaptability of the tracking process by fusing the color features and spatial distribution information of the face.

3 Research methodology

3.1 SMCF technique

3.1.1 Fundamentals

SMCF is a Bayesian estimation through nonlinear and nonparametric Monte Carlo simulation, using samples rather than a description of probability density through functional form. The basic idea is to obtain an optimal estimate of the state using a weighted sum of a series of random samples to represent the desired posterior probability density, and the samples used are one possible description of the target state. When the number of samples increases to infinity, the Monte Carlo simulation property will be equivalent to the posterior probability density function representation, so that the filtering accuracy approximates the optimal estimation. The description of it in the mathematical context is as follows.

Let { X 0 : k i , w k i } be the requested posterior probability density p ( X 0 : k y 1 : k ) of a random observation, and let { X 0 : k i , i = 0 , , N s } be a set of support points. The corresponding weight to the point set is { w k i , i = 0 , , N s } . X 0 : k i = { x j , j = 0 , , N s } denotes the set of states at the time 0 k with normalized weights w i i w k i = 1 . Thus, the posterior probability density at the moment of k can be approximated as Eq. (1).

(1) p ( x 0 : k y 1 : k ) i = 1 N s w k i δ ( x 0 : k x 0 : k i ) .

The results can be approximated with an accuracy guaranteed by the theory of strong number laws. The outstanding advantage of the above method is that it is no longer limited by linear and Gaussian distributions and is in principle applicable to any nonlinear system that can be represented by a state space model. In addition, the method is also capable of parallel processing because of the independent homogeneous distribution of the collected samples. Compared with other forms of recursive Bayesian filters, the method is also flexible, easy to implement, and has high accuracy of approximation results.

3.1.2 SMCF technology implementation framework

Particle filtering is a practical algorithm for solving Bayesian probability, also known as condensation algorithm: bootstrap filtering, interacting particle approximation, sequential Monte Carlo methods, and so on. A particle is a filter of very small scale, which can be thought of as a point representing all possible states of the target. Filtering refers to the current state of the target that can be “filtered,” and in estimation theory, it also refers to the current state of the target estimated by the current and previous observations. The meaning of the particle filter is that the posterior probability of the propagation of the target state can be approximated by several particles. The particle filter can be applied to any nonlinear system which can be represented by state space model and the nonlinear system which cannot be represented by traditional Kalman filter. The accuracy can approximate the optimal estimation. The particle filter method is very flexible, easy to implement, has a parallel structure, and is quite practical. The computational load of recursive Bayesian filtering based on Monte Carlo simulation method is greater than that of Kalman filter and other mathematical equation solving forms. However, particle filter can be implemented in parallel. With the increase in computer speed and the emergence of parallel computers, particle filter has more practical value than traditional Bayesian filters (Kalman filter and mesh filter). By analyzing the tracking problem, a more efficient and faster particle filter algorithm can be constructed.

The SMCF algorithm typically utilizes multiple features to achieve target tracking. These features can be divided into the following categories:

  1. Color features: These features typically include color information of the target area, such as color histograms, color channel histograms, etc. Color features are very useful for tracking color targets and can help algorithms distinguish between targets and backgrounds.

  2. Texture features: Texture features describe the texture information of the target area, such as grayscale co-occurrence matrix, LBP, Gabor filter response, etc. These features help the algorithm recognize the texture characteristics of the target surface.

  3. Spatial features: Spatial features describe the position and shape information of the target in the image. This includes the center position, scale, direction, etc., of the target. These features can be used to determine the position and attitude of the target.

  4. Deep features: Deep features typically come from the middle layer of CNN or other deep learning models. These features can extract high-level semantic information of the target, such as shape, texture, color, etc.

The SMCF algorithm usually comprehensively utilizes these different types of features to improve the performance of target tracking through feature fusion or feature selection. This multi feature fusion method helps to overcome challenges in different scenarios and conditions, making the algorithm more adaptable and robust. Feature engineering and feature selection are important components of the SMCF algorithm to ensure the selection and utilization of the most informative features for efficient target tracking. Based on the above discussion and analysis of the basic principles of SMCF technology, the basic filtering steps can be summarized as follows:

  1. Initial sampling: The initial moment is k = 0 , according to the prior distribution p ( x 0 ) , the initial state sample set { X K i } i = 0 , , N p ( x k ) is established.

  2. Importance sampling: At the moment k > 0 according to the expected distribution q ( x k x k 1 j , z k ) and the collected sample X K i that makes X K i q ( x k x k 1 j , z k ) .

  3. Calculation and normalization of weights. The weights of each sample are calculated and normalized using Eqs. (2)−(4).

    (2) w k i p ( z k x k i ) p ( x k i x k 1 i ) q ( x k x 0 : k 1 i , z 1 : k ) ,

    (3) w = i = 1 N w k i ,

    (4) w k i = w k i w .

  4. Resampling: According to the formula N eff = 1 i = 1 N w k i 2 , find out N eff , and if N eff N th , go to step 2, otherwise, execute the resampling strategy to get a new sample set { X K i , 1 / N } { x k i , w k i } .

  5. Let k = k + 1 , return to step 2.

3.2 Face tracking algorithm

Face tracking is an important task in the field of computer vision, with the goal of detecting and tracking the position and motion of faces in consecutive frames of video. Face tracking typically involves the following key steps:

Face detection: First, it is necessary to detect the presence of faces in each frame. This can be achieved by using face detectors or deep learning models. Some commonly used face detectors include Haar cascade classifiers, SVM classifiers based on HOG features, and deep-learning-based face detection models (such as MTCNN, SSD, YOLO, etc.).

Target initialization: Once a face is detected, the tracker needs to be initialized in order to start tracking. Usually, the detected facial position can be used as the target for the initial tracker.

Target tracking: After initialization, the tracker will update the position of the target in consecutive frames to maintain tracking. Common tracking algorithms include Kalman filtering, mean shift, correlation filters (such as KLT trackers), and deep-learning-based target trackers (such as GOTURN, DeepSORT, etc.).

Occlusion processing: In some cases, the face may be occluded or partially occluded, which requires the tracker to be able to handle occlusion situations. Some trackers have occlusion detection and processing mechanisms, which can adjust the tracking of targets based on the degree of occlusion.

Reinitialization: If the tracker fails in a certain frame (for example, due to occlusion or loss), it needs to be reinitialized, usually using a face detector to redetect the face and restart tracking.

Performance evaluation: During the tracking process, different performance evaluation indicators can be used to evaluate the performance of the tracker, such as accuracy, robustness, speed, etc.

Multi-target tracking: In some scenarios, it is necessary to track multiple faces simultaneously, which is called multi-target tracking. Multi-target tracking algorithms can simultaneously track multiple targets and manage their relationships.

3.2.1 Sample construction and training

In this study, the sample training and construction of target tracking are done through online update. Most discriminative tracking algorithms use random sparse sampling for samples, which can neither extract effective features nor lead to a large number of overlapping samples and too few samples. In this study, a series of image blocks obtained by circular offset matrix are used as training samples, and the problem solution is converted to Fourier transform domain by the characteristics of circular matrix, which can save the sample extraction time by avoiding the matrix inversion operation [8].

As an example, a one-dimensional target vector is represented by an n*1 vector of tracking targets, denoted as x = ( x 1 , x 2 , x 3 , , x n ) , as in Eq. (5).

(5) P = 0 0 0 1 1 0 0 0 0 1 0 0 0 0 1 0 .

The set of samples is obtained from the base sample x and the circular offset X = { P u x u = 0 , , n 1 } .

For two-dimensional images, the training samples can be constructed by cyclically shifting the horizontal and vertical coordinates respectively, thus having x generate a cyclic matrix, as in Eq. (6).

(6) X = C ( x ) = x 1 x 2 x 3 x n x n x 1 x 2 x n 1 x 2 x 3 x 4 x 1 .

In Eq. (6), the first row is the basic sample and the whole matrix is all samples, obtained by circular shifting of the basic samples.

Sample training is a kernel ridge regression problem, which is actually a modification of the least squares method. The ultimate goal of training is to find a regression function that can make predictions on the sample. In constructing the sample, round-robin shift sampling is performed around the target region, and then a sufficient number of samples are generated while obtaining a computationally fast model [9,10].

Linear regression is generally expressed as f ( w , x ) = w T x , treating the sample training problem as a minimized regularized risk optimization problem, as in Eq. (7).

(7) min w i n ( f ( w , x i ) y i ) 2 + λ w 2 ,

where λ is the regularization parameter, which is used to avoid overfitting. In this study, we take λ = 0.01 , and y i is the sample label, which is Gaussian distributed at the center of the target to ensure the maximum probability at the center.

The solution of Eq. (7) is given by

(8) w = ( X H X + λ I ) 1 X H Y .

The circular matrix can be diagonalized by the discrete Fourier transform matrix as in Eq. (9).

(9) X = C ( x ) = F H diag ( x ˆ ) F .

In Eq. (9), C ( x ) is the circular shift function.

Eq. (8) can be changed as follows:

(10) w = x ˆ * y ˆ x ˆ * x ˆ + λ .

Introducing the kernel function ψ(x), w is expressed as a linear combination of x for the sample of Eq. (11).

(11) w = i = 1 n α i ψ ( x i ) .

Let the product between different samples be the result of the combination of kernel functions, as in Eq. (12).

(12) K ij = k ( x i , x j ) .

The final regression function becomes Eq. (13).

(13) f ( z ) = w T z = i = 1 n α i k ( z , x i ) .

Therefore, the solution of the kernel function-based ridge regression is Eq. (14).

(14) a = ( K + λ I ) 1 Y ,

where K denotes the kernel function cycle matrix, K = C ( k xx ) , and k xx is the first row of the kernel matrix K, as shown in the Gaussian kernel function Eq. (15). Gaussian kernel function, also known as radial basis function, is a kernel function used in machine learning algorithms such as SVM. It is very useful in mapping data from input space to high-dimensional feature space and is commonly used for nonlinear classification and regression tasks.

(15) k xx = exp 1 σ 2 ( || x | | 2 + || x | | 2 2 F 1 ( x ˆ c * x c ) ) .

Eq. (15) is called kernel correlation filtering, where σ 2 is the dot product operation between elements, F 1 is the Fourier inverse transform, and the solution of α in the frequency domain is α ˆ = y ˆ k ˆ zz + λ .

3.2.2 Location estimation

In face tracking, it is necessary to determine the position of the face in the next frame based on the information of the given previous frame. The response map is obtained by convolving the trained template with the candidate region as shown in Figure 1, and the response value with the largest value is the center of the face [11]. Convolutional operation is a commonly used technique in deep learning to extract feature information from input data. In object detection and image segmentation tasks, CNNs are typically used to generate response maps to locate objects or bounding boxes of interest. The following describes the process of obtaining a response map by convolving the trained template with candidate regions:

  1. Select template (convolutional kernel): First, you need to define a template for detecting targets or specific features. This template is usually a small two-dimensional weight matrix, which can be manually designed or learned through training. The size and shape of the template are usually related to the target or task to be detected.

  2. Prepare candidate regions: For target detection or image segmentation tasks, a set of candidate regions needs to be generated. These candidate regions are usually rectangular or non-rectangular regions on the image, which may contain objects of interest. Candidate regions can be generated through different methods, such as selective search or region based CNNs.

  3. Convolution operation: For each candidate region, perform a convolution operation on the template and that region. The convolution operation actually involves multiplying the pixel values of the template and candidate regions by each element, and then summing the results.

  4. Response map post-processing: After obtaining the response map, post-processing steps are usually required, such as applying non-maximum suppression to eliminate overlapping candidate regions, and setting thresholds to determine the final detection or segmentation results.

Figure 1 
                     Schematic diagram of face tracking.
Figure 1

Schematic diagram of face tracking.

This process allows neural networks to locate regions or features of interest in the image and generate response maps, where pixel values represent the degree of template matching at different positions. This is one of the commonly used techniques in object detection and image segmentation tasks, which can help machine vision systems recognize and locate objects and features in images.

In the next frame of the video sequence, the candidate region of the target can be identified, and the set of detection samples z is constructed by circular shifting as well, and the response values of all samples can be represented by Eq. (16).

(16) f ( z ) = K z α ,

where K z = C ( k xz ) is the kernel correlation matrix composed of training and test samples z, and f ( x ) is the response map corresponding to the cyclic matrix samples composed of the candidate region z. Using the convolution property of the cyclic matrix F ( C ( x ) * y ) = x ˆ * y ˆ , Eq. (17) is obtained.

(17) F 1 ( f ( z ) ) = ( k ˆ xz ) * α ˆ = k ˆ xz α ˆ .

From Eq. (17), the confidence response map of the candidate box can be quickly obtained. The position with the largest response value is also the position with the highest probability, i.e., the most likely position of the tracking target.

3.2.3 Retracking strategy

During the learning process of human, many people tend to unconsciously cover part of the face with their hands, which will make the face detected by the tracker often incorrect at this time, when the appearance model is more about the hand. Using the hand information for template update will lead to the failure of subsequent tracking and affect the evaluation of the learning effect. Traditional correlation filtering such as KCF lacks retracking mechanism, so when the face is no longer obscured by the collection, the tracking target will be lost [12].

In the target response map, the location with the highest response value is used as the predicted target. In case of good tracking condition, the target response value will be larger, while in vice versa condition, the response map peak will be smaller. So, the peak response size can be used as a basis for whether the face tracking fails or not.

In this study, the primary side regulated (PSR) is used to determine whether a face is lost to follow. Using the response map peak g max and the mean and variance of the other pixels in the response map excluding the 11 × 11 pixels around the peak μ s 1 and variance σ s 1 . The process is shown in equation (18).

(18) PSR = g max μ s 1 σ s 1 .

From Eq. (18), it is known that PSR calculates the relative value of the peak value and the side value. When the face tracking situation is good, PSR is relatively large and the position where the response peak is located is the new position of the face. Conversely, the smaller the PSR is, the face is in an obscured situation. When the PSR is larger than 20, the face tracking is good, when the PSR is about 7, the face is lost or lost in a large area. When the PSR is larger than 3 and smaller than 10, it can be judged that the face tracking result is not reliable at this time.

If there is a PSR calculation such that the target is in an occluded or lost state, then the target model and template are not updated and the target needs to be re-detected.

In this study, an inverted pyramid is used to search for candidate boxes.

In tracking, the time of two adjacent frames is short, so they have continuity in time space. The small displacement of the target in these two frames indicates that the target in the current frame must be in the vicinity of the target position in the previous frame. In the case of consecutive occlusion of multiple frames, the candidate frame search range should be expanded. Regions 1, 2, and 3 do not overlap each other and are rectangular rings with the same center point. The distance from regions 1, 2, and 3 to the center increases one at a time, and the search range increases in turn. Each candidate frame is convolved with the template and the frame with PSR greater than a certain threshold is selected as the new position of the target where the current frame is located [13,14].

3.2.4 Main flow of the algorithm

The main process of the algorithm is summarized as follows.

In order to verify the feasibility and effectiveness of the proposed method, simulation experiments are carried out, respectively, under the condition of face rotation and size change, and the particle filter algorithm based on color histogram is compared with the proposed algorithm.

Input: Image sequence I t , target position P t 1 , target scale S t 1 , position estimation model α t 1 , scale estimation model.

Output: Target position P t and target scale S t .

Step 1. For the first frame, the target features are first extracted, while using the circular matrix to construct training samples with 2.5 times the region of the real target scale. Sample labels y and y s are Gaussian distributed and y s is an N-dimensional vector, and the template is trained and saved.

Step 2. For t frame, a candidate frame is constructed with P t 1 . The response map is obtained by convolving with the template, and the one with the largest response value is used as the position of the target in the current frame, and the PSR is calculated.

Step 3. If the tracking result is unreliable, the inverted pyramid method is used to search several candidate frames and correlate them with the template, and the candidate frame with PSR greater than a set threshold is selected as the target position [15].

Step 4. If the tracking result is reliable, after acquiring the new position of the target in the current frame, please extract the features and construct a feature pyramid with S t 1 , then use the scale filter to perform the correlation operation and get the scale with the largest response value as the target scale. The scale with the largest response value is used as the target scale S t .

Step 5. According to the obtained P t and S t , the training samples are constructed with the circular matrix and the model is updated.

4 Analysis of results

In this study, the widely-used 50 video sequences from benchmark is used, which gives true annotation for each video sequence, divided into 11 attributes, including illumination change, occlusion, scale change, etc. The configuration of this experimental platform is Intel(R) Core(TM) i5-7300HQ, 2.5 GHz, 8GB RAM, windows 10 [16]. Improving the accuracy of the three attributes of proportional change, occlusion, and out of view usually requires a series of improvement measures to enhance the ability of target tracking algorithms to cope with these complex situations. Here are some methods:

  1. Multi scale feature extraction:

    In order to better handle scale changes, multi-scale feature extraction methods can be used. This means extracting features from image pyramids at different scales, so that the algorithm can track targets at different scales. This can improve the scale invariance of the target, making the algorithm more robust.

  2. Occlusion detection and processing:

    To cope with occlusion situations, an occlusion detection module can be introduced to detect whether the target is obstructed by other objects. When occlusion occurs, the algorithm can adopt corresponding strategies, such as predicting the position of the target or attempting to reposition the target.

  3. Long term tracking memory mechanism:

To handle situations where the target is out of sight, a long-term tracking memory mechanism can be introduced. This means that the algorithm can remember the features of the target when it reappears in the field of view, in order to retrace faster. By taking these improvement measures comprehensively, the accuracy of attributes such as proportional changes, occlusion, and out of view can be improved, making the target tracking algorithm more robust and able to maintain high accuracy in more complex situations. This is very important for target tracking tasks in practical applications, especially in situations involving variable attributes.

4. Evaluation indicators:

Distance precision is the percentage of the total number of frames within the threshold distance of the predicted position to the given accuracy value, which is the average Euclidean distance between the center of the tracking target and the center of the real target. The larger the value, the higher the success rate of tracking.

The success rate map is the degree of overlap between the target tracking frame and the true position, as shown in Eq. (19), where R a R b and R a R b are the predicted bounding box and the true bounding box, respectively.

(19) score = area ( R a R b ) area ( R a R b ) ,

where score is the overlap rate between the tracked target and the actual target in each frame, R a is the rectangular area of the target obtained in the current frame in the experiment, R b is the actual annotation of the current frame, area denotes the area of the region, and the intersection of R a and R b areas are divided by their union, score ∈ [0,1]. OP is defined as the percentage of frames with score greater than a certain threshold to the total number of frames. The larger the value, the higher the accuracy. In this study, the threshold value is set to 0.5.

5. Overall performance evaluation:

The overall performance of the five trackers is evaluated with 50 standard video sequences. From Figures 2 and 3, it can be seen that the algorithm in this study outperforms the basic KCF in terms of evaluation accuracy and success rate, and outperforms other tracker algorithms in benchmark, achieving scores of 0.837 and 0.705, respectively. In terms of overlapping accuracy, the reason why the algorithm in this study is higher than KCF is that this study determines the current target tracking status by calculating the PSR value when the target is obscured or lost, so that the tracking error does not accumulate.

Figure 2 
               One-time evaluation of accuracy.
Figure 2

One-time evaluation of accuracy.

Figure 3 
               One-time evaluation of the success rate.
Figure 3

One-time evaluation of the success rate.

6. Evaluation based on different video attributes:

As shown in Figure 4(a)−(e), the tracking algorithm in this study is not ranked first in the two attributes of motion blur and low resolution, but the rankings of the other nine attributes belong to the first place. Compared with the KCF algorithm, the accuracy plots for the three attributes of scale change, occlusion, and leaving the field of view are improved by 10.26, 13.48, and 13.04%, respectively [17].

Figure 4 
               Accuracy graph of five trackers.
Figure 4

Accuracy graph of five trackers.

The algorithms run at 94.7 feet per second (fps) for SMCF, 176.6 fps for KCF, 26.5 fps for Struck, 55.6 fps for TLD, and 67.8 fps for DSST. The comparison results in terms of algorithm running speed show that the SMCF algorithm runs at a speed of 94.7 fps, which is indeed slightly slower compared to the 176.6 fps of the KCF algorithm. This performance difference may be due to the use of feature fusion in the SMCF algorithm, which introduces additional computational complexity. However, although the speed is slightly slower compared to the KCF algorithm, the SMCF algorithm can still meet the real-time requirements of the system [18].

In the presence of attributes such as occlusion and scale transformation in the video set, the SMCF algorithm proposed in this study performs well in accuracy and success rate graphs, and has better performance compared to other algorithms. This indicates that the SMCF algorithm has excellent robustness and performance in dealing with occlusion and scale issues. Specifically, accuracy maps and success rate maps are important indicators for evaluating the performance of target tracking algorithms. From the results of these two graphs, it can be seen that the SMCF algorithm exhibits higher accuracy and success rate in handling occlusion situations in videos. This means that the SMCF algorithm can more accurately retrace the target when the occluded part reappears, and also better adapt and maintain stable tracking performance when the target scale changes. This discovery emphasizes the excellent performance of SMCF algorithm in complex video scenes, especially in the face of challenges such as occlusion and scale transformation, which can effectively improve tracking accuracy and robustness. Therefore, the SMCF algorithm has higher feasibility and practicality in solving practical video analysis problems [19,20].

4.1 Discussion

Video target tracking has become the focus of research gradually, and it has important applications in military, civilian, and many other fields. In this study, the tracking algorithm of video moving object is studied. First, the significance and purpose of video tracking are expounded, the development trend of video tracking at home and abroad is introduced, and the algorithm of video target tracking is summarized. In Section 2, we study the most basic Kalman filtering algorithm and two improved Kalman filtering algorithms, namely, the EKF algorithm and the unscented Kalman filtering (UKF) algorithm, analyze their advantages and disadvantages, and carry out state simulation experiments for the above algorithms. In Section 3, the particle filter algorithm is mainly studied. Particle filter is a filter algorithm based on Bayesian estimation theory and Monte Carlo method, which can be applied in nonlinear model and non-Gaussian noise environment. In this section, the problems of particle weight degradation in this algorithm are analyzed, and two solutions are proposed, namely, optimization of importance density function and resampling. In Section 4, the ideas of Kalman filter and particle filter are combined, and the unscented particle filter (UPF) algorithm based on minimum deviation sampling is proposed. The important distribution function of the particle filter is generated by UKF, which makes great use of the latest observation data and overcomes the weight degradation problem effectively. Finally, this work studies the target tracking method based on video sequence, improves the video tracking algorithm based on color information, and proposes a video target tracking algorithm based on multi-feature fusion. The algorithm takes SIR Particle filter as the tracking framework, and comprehensively considers the color information and motion information of the target when calculating the particle weight. In this way, the discrete particle set obtained can more truly simulate the posterior probability distribution of the state vector, so that more stable and accurate tracking effect can be obtained.

5 Conclusion

In this study, the Kalman filter algorithm and the particle filter algorithm are combined, and the importance distribution of the particle filter is generated by UKF, which makes use of the latest observation data and greatly reduces the estimation error. Experiments show that the proposed UPF algorithm based on minimum skewness sampling and partition resampling not only achieves high accuracy, but also has good stability performance. At the same time, the operation efficiency of the algorithm is also high. In addition, the object tracking method based on video sequence is also introduced. By analyzing the existing classical algorithms, an improved target tracking algorithm is proposed. The proposed algorithm takes SIR particle filter as the tracking framework, and comprehensively considers the color information and motion information of the target when calculating the particle weight. In this way, the discrete particle set obtained can more truly simulate the posterior probability distribution of the state vector, so that more stable and accurate tracking effect can be obtained.

The unscented transformation strategies and particle filter resampling strategies are researched, and a scented algorithm based on minimum deviation sampling (UPF) is proposed, and the scented strategy is applied to the UKF algorithm. The proposed distribution is generated and sampled by UKF method, which solves the problem of particle degradation caused by the conversion of prior density function as the alternative distribution in the general particle filter algorithm. Finally, this study proposes a new video target tracking algorithm, which uses particle filter as a framework to integrate the color information and motion information in the observed data. Theoretical analysis and experimental results show that the improved UPF algorithm improves the stability and precision of the filter, and the operation efficiency of the algorithm is also increased by 3. At the same time, the new video target tracking algorithm effectively improves the robustness and tracking accuracy of the algorithm.

For the principle that the object of learning effect evaluation remains unchanged during online learning, a face tracking algorithm based on correlation filtering is proposed, which is simple, effective, easy to operate, and can solve the problems of scale change and occlusion during online learning tracking, so as to ensure the smooth implementation of learning effect evaluation. The results show that the algorithm in this study outperforms the basic KCF in terms of evaluation accuracy and success rate, as well as outperforms other tracker algorithms in benchmark, achieving scores of 0.837 and 0.705, respectively. The algorithm in this study is validated on the OTB database, which shows that the improved algorithm is effective. This study also has some limitations and requires further research to improve. First, the methods in the study cannot solve all the problems in the corresponding field, but rather have a certain degree of pertinence and applicability. Due to the complexity of facial information processing, specific algorithms can only be applied to specific applications and solve specific problems. When constructing practical application systems, multiple factors must be considered to enhance the robustness of the algorithm, but this inevitably increases the computational complexity. Therefore, research on facial detection, tracking, recognition, and understanding based on computer vision is far from reaching the level of application. The algorithm of this study is mainly based on the characteristics of skin color, how to combine other features of the face, how to explore the clues of voice or video text title to further improve the robustness of tracking is the direction of the next research.

Acknowledgements

This work was supported by the Basic Scientific Research Operating Expenses of Heilongjiang Provincial Universities and Colleges under Grant 2018-KYYWF-0940.

  1. Author contributions: Each author made significant individual contributions to this manuscript. Yunming Du: writing and performing surgeries; Jing Tia: data analysis and performing surgeries; Yi Liu: article review and intellectual concept of the article.

  2. Conflict of interest: The authors declare that they have no competing interests.

  3. Data availability statement: The data used to support the findings of this study are available from the corresponding author upon request.

References

[1] Daníş FS, Cemgil AT, Ersoy C. Adaptive sequential Monte Carlo filter for indoor positioning and tracking with bluetooth low energy beacons. IEEE Access. 2021;9:37022–38.10.1109/ACCESS.2021.3062818Suche in Google Scholar

[2] Dai C, Heng J, Jacob PE, Whiteley N. An invitation to sequential Monte Carlo samplers. J Am Stat Assoc. 2022;117(539):1587–600.10.1080/01621459.2022.2087659Suche in Google Scholar

[3] Khazeiynasab SR, Qi J. Generator parameter calibration by adaptive approximate Bayesian computation with sequential Monte Carlo sampler. IEEE Trans Smart Grid. 2021;12(5):4327–38.10.1109/TSG.2021.3077734Suche in Google Scholar

[4] Suruzhon M, Bodnarchuk MS, Ciancetta A, Wall ID, Essex JW. Enhancing ligand and protein sampling using sequential Monte Carlo. J Chem Theory Comput. 2022;18(6):3894–910.10.1021/acs.jctc.1c01198Suche in Google Scholar PubMed PubMed Central

[5] Yamamoto Y, Yajima T, Kawajiri Y. Uncertainty quantification for chromatography model parameters by Bayesian inference using sequential Monte Carlo method. Chem Eng Res Des. 2021;175:223–37.10.1016/j.cherd.2021.09.003Suche in Google Scholar

[6] Du Q, Guyader A. Variance estimation in adaptive sequential Monte Carlo. Ann Appl Probab. 2021;31(3):1021–60.10.1214/20-AAP1611Suche in Google Scholar

[7] Gloaguen P, Le Corff S, Olsson J. A pseudo-marginal sequential Monte Carlo online smoothing algorithm. Bernoulli. 2022;28(4):2606–33.10.3150/21-BEJ1431Suche in Google Scholar

[8] Wee DCH, Chen F, Dunsmuir WT. Likelihood inference for Markov switching GARCH (1, 1) models using sequential Monte Carlo. Econ Stat. 2022;21:50–68.10.1016/j.ecosta.2020.03.004Suche in Google Scholar

[9] Banerjee S, Singh GK. Monte Carlo filter-based motion artifact removal from electrocardiogram signal for real-time telecardiology system. IEEE Trans Instrum Meas. 2021;70:1–10.10.1109/TIM.2021.3102737Suche in Google Scholar

[10] Li D, Clements A, Drovandi C. Efficient Bayesian estimation for GARCH-type models via Sequential Monte Carlo. Econ Stat. 2021;19:22–46.10.1016/j.ecosta.2020.02.002Suche in Google Scholar

[11] Martino L, Elvira V. Compressed Monte Carlo with application in particle filtering. Inf Sci. 2021;553:331–52.10.1016/j.ins.2020.10.022Suche in Google Scholar

[12] Bon JJ, Lee A, Drovandi C. Accelerating sequential Monte Carlo with surrogate likelihoods. Stat Comput. 2021;31:1–26.10.1007/s11222-021-10036-4Suche in Google Scholar

[13] Buchanan R, Bednarek J, Camurri M, Nowicki MR, Walas K, Fallon M. Navigating by touch: haptic Monte Carlo localization via geometric sensing and terrain classification. Auton Robot. 2021;45(6):843–57.10.1007/s10514-021-10013-wSuche in Google Scholar

[14] Xu J. A deep learning approach to building an intelligent video surveillance system. Multimed Tools Appl. 2021;80(4):5495–515.10.1007/s11042-020-09964-6Suche in Google Scholar

[15] Gothane S. A practice for object detection using YOLO algorithm. Int J Scientif Res Comput Sci Eng Inf Technol. 2021;7(2):268–72.10.32628/CSEIT217249Suche in Google Scholar

[16] Zheng G, Xu Y. Efficient face detection and tracking in video sequences based on deep learning. Inf Sci. 2021;568:265–85.10.1016/j.ins.2021.03.027Suche in Google Scholar

[17] Badawi H, Shawagfeh N and Arqub OA. Fractional conformable stochastic integrodifferential equations: existence, uniqueness, and numerical simulations utilizing the shifted Legendre spectral collocation algorithm. Math Probl Eng. 2022;80:3981–96.10.1155/2022/5104350Suche in Google Scholar

[18] Arqub OA. The reproducing kernel algorithm for handling differential algebraic systems of ordinary differential equations. Math Methods Appl Sci. 2016;2:235–45.10.1002/mma.3884Suche in Google Scholar

[19] Momani S, Maayah B, Arqub OA. The reproducing Kernel algorithm for numerical solution of van der Pol damping model in view of the Atangana–Baleanu fractional approach. Personal Ubiquitous Comput. 2021;25:129–40.Suche in Google Scholar

[20] Arqub OA. Computational algorithm for solving singular Fredholm time-fractional partial integrodifferential equations with error estimates. J Appl Maths Comput. 2019;59:227–43.10.1007/s12190-018-1176-xSuche in Google Scholar

Received: 2023-05-02
Revised: 2023-09-08
Accepted: 2023-09-10
Published Online: 2023-10-23

© 2023 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Artikel in diesem Heft

  1. Research Articles
  2. The regularization of spectral methods for hyperbolic Volterra integrodifferential equations with fractional power elliptic operator
  3. Analytical and numerical study for the generalized q-deformed sinh-Gordon equation
  4. Dynamics and attitude control of space-based synthetic aperture radar
  5. A new optimal multistep optimal homotopy asymptotic method to solve nonlinear system of two biological species
  6. Dynamical aspects of transient electro-osmotic flow of Burgers' fluid with zeta potential in cylindrical tube
  7. Self-optimization examination system based on improved particle swarm optimization
  8. Overlapping grid SQLM for third-grade modified nanofluid flow deformed by porous stretchable/shrinkable Riga plate
  9. Research on indoor localization algorithm based on time unsynchronization
  10. Performance evaluation and optimization of fixture adapter for oil drilling top drives
  11. Nonlinear adaptive sliding mode control with application to quadcopters
  12. Numerical simulation of Burgers’ equations via quartic HB-spline DQM
  13. Bond performance between recycled concrete and steel bar after high temperature
  14. Deformable Laplace transform and its applications
  15. A comparative study for the numerical approximation of 1D and 2D hyperbolic telegraph equations with UAT and UAH tension B-spline DQM
  16. Numerical approximations of CNLS equations via UAH tension B-spline DQM
  17. Nonlinear numerical simulation of bond performance between recycled concrete and corroded steel bars
  18. An iterative approach using Sawi transform for fractional telegraph equation in diversified dimensions
  19. Investigation of magnetized convection for second-grade nanofluids via Prabhakar differentiation
  20. Influence of the blade size on the dynamic characteristic damage identification of wind turbine blades
  21. Cilia and electroosmosis induced double diffusive transport of hybrid nanofluids through microchannel and entropy analysis
  22. Semi-analytical approximation of time-fractional telegraph equation via natural transform in Caputo derivative
  23. Analytical solutions of fractional couple stress fluid flow for an engineering problem
  24. Simulations of fractional time-derivative against proportional time-delay for solving and investigating the generalized perturbed-KdV equation
  25. Pricing weather derivatives in an uncertain environment
  26. Variational principles for a double Rayleigh beam system undergoing vibrations and connected by a nonlinear Winkler–Pasternak elastic layer
  27. Novel soliton structures of truncated M-fractional (4+1)-dim Fokas wave model
  28. Safety decision analysis of collapse accident based on “accident tree–analytic hierarchy process”
  29. Derivation of septic B-spline function in n-dimensional to solve n-dimensional partial differential equations
  30. Development of a gray box system identification model to estimate the parameters affecting traffic accidents
  31. Homotopy analysis method for discrete quasi-reversibility mollification method of nonhomogeneous backward heat conduction problem
  32. New kink-periodic and convex–concave-periodic solutions to the modified regularized long wave equation by means of modified rational trigonometric–hyperbolic functions
  33. Explicit Chebyshev Petrov–Galerkin scheme for time-fractional fourth-order uniform Euler–Bernoulli pinned–pinned beam equation
  34. NASA DART mission: A preliminary mathematical dynamical model and its nonlinear circuit emulation
  35. Nonlinear dynamic responses of ballasted railway tracks using concrete sleepers incorporated with reinforced fibres and pre-treated crumb rubber
  36. Two-component excitation governance of giant wave clusters with the partially nonlocal nonlinearity
  37. Bifurcation analysis and control of the valve-controlled hydraulic cylinder system
  38. Engineering fault intelligent monitoring system based on Internet of Things and GIS
  39. Traveling wave solutions of the generalized scale-invariant analog of the KdV equation by tanh–coth method
  40. Electric vehicle wireless charging system for the foreign object detection with the inducted coil with magnetic field variation
  41. Dynamical structures of wave front to the fractional generalized equal width-Burgers model via two analytic schemes: Effects of parameters and fractionality
  42. Theoretical and numerical analysis of nonlinear Boussinesq equation under fractal fractional derivative
  43. Research on the artificial control method of the gas nuclei spectrum in the small-scale experimental pool under atmospheric pressure
  44. Mathematical analysis of the transmission dynamics of viral infection with effective control policies via fractional derivative
  45. On duality principles and related convex dual formulations suitable for local and global non-convex variational optimization
  46. Study on the breaking characteristics of glass-like brittle materials
  47. The construction and development of economic education model in universities based on the spatial Durbin model
  48. Homoclinic breather, periodic wave, lump solution, and M-shaped rational solutions for cold bosonic atoms in a zig-zag optical lattice
  49. Fractional insights into Zika virus transmission: Exploring preventive measures from a dynamical perspective
  50. Rapid Communication
  51. Influence of joint flexibility on buckling analysis of free–free beams
  52. Special Issue: Recent trends and emergence of technology in nonlinear engineering and its applications - Part II
  53. Research on optimization of crane fault predictive control system based on data mining
  54. Nonlinear computer image scene and target information extraction based on big data technology
  55. Nonlinear analysis and processing of software development data under Internet of things monitoring system
  56. Nonlinear remote monitoring system of manipulator based on network communication technology
  57. Nonlinear bridge deflection monitoring and prediction system based on network communication
  58. Cross-modal multi-label image classification modeling and recognition based on nonlinear
  59. Application of nonlinear clustering optimization algorithm in web data mining of cloud computing
  60. Optimization of information acquisition security of broadband carrier communication based on linear equation
  61. A review of tiger conservation studies using nonlinear trajectory: A telemetry data approach
  62. Multiwireless sensors for electrical measurement based on nonlinear improved data fusion algorithm
  63. Realization of optimization design of electromechanical integration PLC program system based on 3D model
  64. Research on nonlinear tracking and evaluation of sports 3D vision action
  65. Analysis of bridge vibration response for identification of bridge damage using BP neural network
  66. Numerical analysis of vibration response of elastic tube bundle of heat exchanger based on fluid structure coupling analysis
  67. Establishment of nonlinear network security situational awareness model based on random forest under the background of big data
  68. Research and implementation of non-linear management and monitoring system for classified information network
  69. Study of time-fractional delayed differential equations via new integral transform-based variation iteration technique
  70. Exhaustive study on post effect processing of 3D image based on nonlinear digital watermarking algorithm
  71. A versatile dynamic noise control framework based on computer simulation and modeling
  72. A novel hybrid ensemble convolutional neural network for face recognition by optimizing hyperparameters
  73. Numerical analysis of uneven settlement of highway subgrade based on nonlinear algorithm
  74. Experimental design and data analysis and optimization of mechanical condition diagnosis for transformer sets
  75. Special Issue: Reliable and Robust Fuzzy Logic Control System for Industry 4.0
  76. Framework for identifying network attacks through packet inspection using machine learning
  77. Convolutional neural network for UAV image processing and navigation in tree plantations based on deep learning
  78. Analysis of multimedia technology and mobile learning in English teaching in colleges and universities
  79. A deep learning-based mathematical modeling strategy for classifying musical genres in musical industry
  80. An effective framework to improve the managerial activities in global software development
  81. Simulation of three-dimensional temperature field in high-frequency welding based on nonlinear finite element method
  82. Multi-objective optimization model of transmission error of nonlinear dynamic load of double helical gears
  83. Fault diagnosis of electrical equipment based on virtual simulation technology
  84. Application of fractional-order nonlinear equations in coordinated control of multi-agent systems
  85. Research on railroad locomotive driving safety assistance technology based on electromechanical coupling analysis
  86. Risk assessment of computer network information using a proposed approach: Fuzzy hierarchical reasoning model based on scientific inversion parallel programming
  87. Special Issue: Dynamic Engineering and Control Methods for the Nonlinear Systems - Part I
  88. The application of iterative hard threshold algorithm based on nonlinear optimal compression sensing and electronic information technology in the field of automatic control
  89. Equilibrium stability of dynamic duopoly Cournot game under heterogeneous strategies, asymmetric information, and one-way R&D spillovers
  90. Mathematical prediction model construction of network packet loss rate and nonlinear mapping user experience under the Internet of Things
  91. Target recognition and detection system based on sensor and nonlinear machine vision fusion
  92. Risk analysis of bridge ship collision based on AIS data model and nonlinear finite element
  93. Video face target detection and tracking algorithm based on nonlinear sequence Monte Carlo filtering technique
  94. Adaptive fuzzy extended state observer for a class of nonlinear systems with output constraint
Heruntergeladen am 4.11.2025 von https://www.degruyterbrill.com/document/doi/10.1515/nleng-2022-0329/html?lang=de
Button zum nach oben scrollen