Home Technology Seismic vulnerability signal analysis of low tower cable-stayed bridges method based on convolutional attention network
Article Open Access

Seismic vulnerability signal analysis of low tower cable-stayed bridges method based on convolutional attention network

  • Dingbo Chen EMAIL logo and Liangjun Lai
Published/Copyright: January 9, 2024
Become an author with De Gruyter Brill

Abstract

Due to the particularity and complexity of sedimentary environments, the wave impedance differences between different reflection interfaces in underground media may vary greatly. Therefore, an encoder–decoder neural network is proposed to enhance erroneous seismic weak reflection signals. The convolutional neural network (CNN) has the problem of difficulty in parallel computing, resulting in slow network training and computational efficiency. Considering that attention has an innate global self-attention mechanism, can compensate for long-term dependency deficiencies, and has the ability to perform parallel computing, which greatly compensates for the shortcomings of CNNs and recurrent neural networks, a seismic impedance inversion method based on convolutional attention networks is proposed. To improve the ability to extract noise, residual structure and convolutional attention module (CBAM) were introduced. The residual structure utilizes residual jump to weaken network degradation and reduce the difficulty of feature mapping. The CBAM uses a mixed attention weight of channel and space, which can enhance features with high correlation and suppress features with low correlation. In the decoder, in order to improve the dimension recovery ability of feature fusion, bilinear interpolation is selected for upsampling. The application results of the model and actual data indicate that this method can effectively enhance the weak reflection signals caused by the formation itself and improve the reservoir identification ability of seismic data.

1 Introduction

At present, there are few reports on how to solve the problem of incomplete measurement of vibration data in earthquake simulation experiments in existing research [1,2,3]. Neural networks have been widely used in fields such as earthquake prediction and active control of structural vibration due to their large-scale parallel processing and distributed information storage capabilities, strong self-learning, association, fault tolerance capabilities, good adaptability and self-organization, and nonlinear fitting systems with multiple inputs and outputs [4]. Neural networks also have the advantages of not needing to establish mathematical models and having the ability to identify and predict unknown systems and are often used as tools for data prediction. Guo et al. [5] used BP neural networks to predict the results of different experimental conditions with a small number of experiments. From this, it can be seen that neural networks have the potential to predict incomplete response data in earthquake simulation experiments.

Seismic weak reflection signal enhancement methods mainly rely on convolutional neural networks (CNNs) for image processing, recurrent neural networks (RNNs) for natural language processing, and new networks formed by the combination of these networks [6]. As a CNN, it has the problem of translation invariance in convolution and only focuses on local information, which makes it difficult for convolutional networks to capture and store long-distance dependency information in sequence processing. However, expanding the convolutional kernel and channel number to capture long-term dependencies can lead to the disappearance of gradients, resulting in the lack of convergence in the entire network training [7]. So CNNs have certain limitations. As an RNN, due to its ability to store effective information in various scattered neurons, it cannot effectively preserve those overly long data dependency problems, resulting in insufficient long-term dependency capture in longer sequence problems. Moreover, RNNs are difficult to parallelize effectively, resulting in slow computational efficiency and wasting time and cost during network training. Remote early warning can calculate the three elements of the earthquake [8] through earthquake location and magnitude calculation methods, then predict the intensity of the area where the seismic wave does not reach through the attenuation relationship of ground motion parameters [9]. This method fully considers the source parameters and predicts the intensity through the attenuation relationship obtained by statistics. The intensity distribution of the entire intensity field can be predicted without using the waveform information of all stations. However, the timeliness of this method is not as good as that of local early warning. There is a large blind area for early warning, and the predicted intensity distribution is generally in accordance with a certain law [10]. The isointensity line is generally circular, while the actual intensity distribution is often uneven, and there are some areas with abnormal intensity. Therefore, predicting through attenuation relationships also has certain limitations.

With the development of artificial intelligence, deep learning technology has been applied to many fields, which has also brought new ideas to the field of exploration geophysics. Among them, the field of seismic data processing has achieved some research results through deep learning, and the attenuation of seismic signal noise has become one of the more active research fields. After introducing CNNs, various noise attenuation methods based on CNN architecture have been proposed. Jiang et al. [11] used denoising convolutional networks to attenuate random noise in seismic data, resulting in better denoising results than traditional methods. Luo and Zhao [12] applied the data augmentation method and utilized a U-shaped CNN (Unet) to effectively suppress random noise. Qiao et al. [13] added residual blocks on the basis of Unet to solve the problem of generalization, which not only achieved the denoising effect but also protected effective signals. Zhang et al. [14] mentioned adaptive learning of seismic signal features and the use of convolutional denoising autoencoders to eliminate strong noise. Margheriti et al. [15] constructed a network containing global context and attention mechanisms by improving the network, which suppressed random noise and retained more local details. Guo et al. [16] constructed an unsupervised residual convolutional network that uses double-layer convolutional blocks to improve computational efficiency and achieve adaptive noise attenuation. This article proposes a CNN (RAUnet) that integrates residual attention mechanism [17], introducing convolutional attention mechanism and residual structure to improve denoising effect and enhance model generalization. The denoising performance of the algorithm is demonstrated through experiments on synthetic and actual seismic signals.

This article proposes the seismic weak reflection signal enhancement method based on convolutional attention network. By performing fractional power operation on the reflection coefficient, a sequence of quasi-reflection coefficients is obtained. The synthetic seismogram and pseudo-synthetic seismogram are calculated with the primitive reflexes coefficient and pseudo-reflection coefficient, and the training sample set is generated. The residual [18] network is then trained to establish a mapping relationship between synthetic seismic records and quasi-synthetic seismic records. Finally, the network is applied to seismic data to enhance weak seismic reflection signals. The method in this article is not to eliminate strong seismic signals but to narrow the relative difference between weak and strong reflection coefficients by constructing a power reflection coefficient model, thereby highlighting weak reflection signals.

2 Methods

2.1 RAUnet denoising method

In order to improve the denoising effect and enhance the generalization of the model, this article constructs a CNN (RAUnet) that integrates a residual attention mechanism [19,20]. The network adopts batch standardization and activation functions with leakage rectification after each convolution layer, adds residual structure and convolution attention modules in the encoder, respectively, and uses bilinear interpolation in the decoder for upsampling [21].

Batch normalization (BN) [22] refers to the standardization of each batch of data at the network layer when training a neural network using the gradient descent method. The position of batch standardization is between the convolutional layer and the nonlinear layer, which is regarded as an interlayer mechanism of the network [23]. It provides a commonly used method for reinitializing parameters in deep networks, which can significantly reduce the issue of update coordination between multi-layer networks, thereby achieving the goal of reducing the offset of covariates within the network [24]. The results after batch standardization can be expressed as follows:

(1) μ B = 1 S k = 1 S a k ,

(2) σ B = 1 S k = 1 S ( a k μ B ) 2 ,

(3) a ˜ k = a k μ B σ B 2 + ε ,

(4) b k = γ a ˜ k + β ,

where a k and b k are input samples [25] and output results for BN [26] layer, respectively; μ B and σ B , respectively, represent the expected and standard deviation of the sample [27]; ε is a constant; a ˜ k is a k regularization; and γ and β are leachable parameters [28].

The activation function is located behind the batch standardization layer and is used to control neurons, which can solve the nonlinear problems encountered in seismic signal processing. The commonly used activation function is mainly the linear rectification activation function (ReLU) [29], but if the transfer variable is less than zero, the neurons in the negative value interval will not be updated, which will lead to the loss of effective information. To avoid gradient disappearance, the linear rectification activation function [30] (leaky ReLU, Figure 1) with leakage is used, and its formula is as follows:

(5) g ( h ) = h , h 0 c h , h 0 ,

where h is the input of the activation function [31] and c is a very small constant. The value of c in this article is set to 0.2.

Figure 1 
                  Leaky ReLU activation function.
Figure 1

Leaky ReLU activation function.

Attention mechanisms [32] can be divided into two categories: channel attention and spatial attention. Channel attention can enhance or weaken different channels for different tasks to determine the importance of each feature channel. Spatial attention believes that not all regions contribute equally to the task, mainly emphasizing areas related to the task. Both types of mechanisms have an enhancing effect on the network, and their effective combination brings greater promotion to the network. Therefore, this article adopts a mixed domain attention mechanism combining channel and space [33] methods, adding learning weights to high-frequency data in the channel domain and local correlations in the spatial domain [34], respectively, to construct a convolutional attention module as shown in Figure 2.

Figure 2 
                  Convolution attention module.
Figure 2

Convolution attention module.

It is generally believed that the deeper the network [35], the more likely it is to obtain a model with better performance, but this is not the case. The practice has shown that for certain large datasets [36], even if the number of network layers continues to deepen, the ideal effect cannot be achieved, which can backfire and lead to network degradation and other problems. In response to the above issues, Wu et al. [37] introduced skip connections, namely the residual learning framework (ResNet), which combines shallow and deep features to form the input for subsequent operations, effectively alleviating the loss of input features and improving the performance of feature recovery [38]. This article adopts a residual structure as shown in Figure 3. The skip stacking method involves the stacking of pixel values in the image. When the dimensions are the same, they can be directly added. If they are different, they need to be converted into dimensions first and then added [39].

Figure 3 
                  Residual structure.
Figure 3

Residual structure.

3 Experiment

3.1 Data selection

The training results of artificial neural network algorithms largely depend on the quality of the training samples. Therefore, when selecting data, it is necessary to consider all situations as much as possible, including sufficient data volume and ensuring the uniformity of data distribution. This article first selected 629 earthquake events recorded by the KiKnet strong earthquake network in Japan from 1999 to 2016 for research, with magnitudes ranging from MS3.7 to MS9.0. Among them, there are more seismic events below magnitude 7, and their distribution is uniform, while seismic events above magnitude 7 are relatively rare. In this regard, this article focuses on the analysis of two earthquakes with a magnitude of more than 8, namely, the earthquake with a magnitude of 8.0 on September 26, 2003, and the earthquake with a magnitude of 9.0 on March 11, 2011. Both of them occurred in the waters of Japan, and the recently triggered stations were 103 and 145 km, respectively. In the analysis and testing, it was found that these two earthquakes were unable to accurately predict the intensity of major earthquakes occurring on land, such as the Wenchuan earthquake, and there may be significant errors in predicting the intensity of nearby instruments. Therefore, it was decided to include strong earthquake data from the Wenchuan earthquake on May 12, 2008, the Jiji earthquake on September 21, 1999, and the Ludian earthquake on August 3, 2014, in the sample. There were a total of 632 seismic events used for testing.

If there are too many data choices and different types, it will increase the difficulty of network training and also increase the difficulty of prediction. If there are too few data choices and insufficient sample types available, the training samples may not be convincing. In the selection of stations, try to choose stations with significant predictive value and exclude the influence of meaningless and abnormal stations. The following points should be considered:

  1. Each station of the KiK-net strong earthquake observation network in Japan is divided into surface stations and underground bedrock stations. This time, all three directional data from underground bedrock stations are used, and the three earthquakes added later are data from surface stations due to limited data sources;

  2. According to the instructions in the “China Earthquake Intensity Scale (GB/T 17742-2020)” [38], stations with intensities less than 1 in actual calculations are excluded;

  3. The propagation path of earthquakes with deeper focal depths is complex and has less impact than shallow earthquakes. This article focuses on predicting shallow earthquakes and excludes earthquake-triggering stations with focal depths greater than 50 km;

  4. Small earthquakes and large earthquakes with larger focal distances are not the objects of early warning research. In record screening, the relationship between magnitude and focal distance was found based on the influence range of the earthquake, as shown in the following equation:

(6) lg R 0.86 + 0.17 M ,

where R is the source distance and M is the magnitude. In the process of data selection, it is also necessary to exclude stations that have experienced head loss, data anomalies, and other situations. According to this rule, a total of 2,231 stations were selected in this article. The distribution of the relationship between the focal distance, magnitude, and intensity of these stations is shown in Figures 9 and 10. The distribution of the selected stations in this article should be as uniform as possible and represent the triggering situation of the vast majority of seismic stations.

Figure 4 
                  Time history curve of an artificial seismic wave. (a) Artificial seismic wave 1.
Figure 4

Time history curve of an artificial seismic wave. (a) Artificial seismic wave 1.

Figure 5 
                  Acceleration curve of measurement point 7 under working condition 2.
Figure 5

Acceleration curve of measurement point 7 under working condition 2.

3.2 Overview of vibration table testing and construction of neural networks

In order to study the dynamic coupling effect between valves and pipelines in nuclear power valve pipeline systems and to discuss the dynamic behavior of valve pipeline systems during earthquakes, a “valve pipeline system seismic simulation test” was designed for pipeline systems without valves and with valves installed [39]. The nodes of the tube unit in the model correspond one-to-one to the positions of the experimental measurement points, and the position number of the measurement points is used to refer to the nodes. Due to the complete symmetry of the pipeline structure in the Y and Z directions, only the Y-direction test was designed for it during the testing process. Condition 1 in Table 1 is to investigate the dynamic characteristics of pipeline structures. In order to study the response of pipeline structures under different seismic excitations and provide experimental data for seismic margin analysis of pipeline structures, two artificial seismic wave tests with different amplitudes were designed, namely condition 2 and condition 3. The maximum amplitude of the artificial seismic wave used in condition 2 is 13.2 m/s2, as shown in Figure 4(a). The maximum amplitude of the artificial seismic wave used in condition 3 is 8.6 m/s2, as shown in Figure 4(b).

Table 1

Pipeline structure test conditions

Test structure Condition No. Duration, s Excitation waveform
Piping system 1 120 White noise
2 60 Y-direction artificial seismic wave 1
3 60 Y-direction artificial seismic wave 2
Figure 6 
                  Relationship between focal distance and magnitude of selected seismic event stations in this article.
Figure 6

Relationship between focal distance and magnitude of selected seismic event stations in this article.

The acceleration time history curve measured at the maximum position measurement point 7 in condition 2 is shown in Figure 5. From Figure 6, it can be seen that in the response time history curve of this measuring point, there are a total of 14 sampling points in the time periods of 17.226–17.228, 17.338–17.344, and 17.399–17.402 s. The same acceleration data obtained from adjacent sampling points clearly does not conform to objective laws, indicating that under the action of artificial seismic wave 1, the response of measurement point 7 may have errors or incomplete measurements during the above time period. Due to the numerous working conditions involved in the entire process and the fact that the test structure was replaced several times during the test, this issue was not discovered during the test process, and the test bench was removed after the completion of the test, making it impossible to repeat the test. In order to make the experimental data available and achieve the experimental objectives, a neural network is used to predict the erroneous data in the experiment to supplement the time history curve of measurement point 7 in condition 2.

Figure 7 
                  Relationship between focal distance and intensity of selected seismic event stations in this article.
Figure 7

Relationship between focal distance and intensity of selected seismic event stations in this article.

Figure 8 
                  Predicted average intensity difference of three models from 1 to 20 s changes in absolute values.
Figure 8

Predicted average intensity difference of three models from 1 to 20 s changes in absolute values.

In order to avoid significant differences between local and global minimum values as much as possible, the calculations for each network are carried out multiple times, starting from multiple randomly selected initial value points. For each model, fitting is performed using parameters ranging from 1 to 20 s, resulting in a total of 60 network models. This article conducts result statistics on each model calculated, using each network to predict the absolute value of the intensity difference between each station in the training sample and confirmation sample and using its average value as the discriminant criterion. The specific results are shown in Figure 7.

Figure 9 
                  Predicted intensity difference of the first and 20th seconds of three models distribution of absolute values. (a) Model 1: distribution of absolute value of preheating intensity difference, (b) Model 2: distribution of absolute value of preheating intensity difference, (c) Model 3: distribution of absolute value of preheating intensity difference.
Figure 9

Predicted intensity difference of the first and 20th seconds of three models distribution of absolute values. (a) Model 1: distribution of absolute value of preheating intensity difference, (b) Model 2: distribution of absolute value of preheating intensity difference, (c) Model 3: distribution of absolute value of preheating intensity difference.

From Figure 8, the following conclusion can be drawn: (i) The predicted results improve with time, and the effect increases faster in the first 15 s. (ii) The prediction result of model 3 is better than that of model 2, and model 2 is better than that of model 1. Model 3 has a good prediction result since the first second, which shows that accurate source distance and magnitude play a great role in improving the prediction of intensity. (iii) The prediction results of the training samples are slightly better than those of the confirmation samples, indicating that the model can also make good predictions for the same type of data that is not used for training. (iv) The predicted results of the three models are all less than 0.6 by 20 s, and in practical applications, most of the intensity differences after rounding the results are 0 or 1. In actual prediction, if the absolute value of intensity difference can be within 1 and the rounded value is also within 1, it is a good and acceptable result in an earthquake warning. In Figure 8, the change in the average value of the intensity difference is analyzed, while the distribution of intensity difference is also analyzed. First, three models were used to conduct a detailed analysis of the average intensity difference between the representative first and 20th seconds of the training samples (Figure 9). The other intermediate moments were not listed one by one, and they were a gradually changing process. From Figure 9, it can be seen that at the first second, the proportion of predicted intensity difference less than 1 using model 3 is significantly higher than the other two models. Model 1 has poor prediction performance, with 7.92% of predicted intensity difference greater than 3. At the 20th second, the prediction performance of the three models was relatively good, and the proportion of predicted intensity differences less than 1 was higher than 90%, with relatively accurate results. Second, this article analyzes the proportion of the prediction intensity difference of the three models from 1 to 20 s for training samples that is less than 1 (Figure 10), which can to some extent reflect the situation of intermediate moments not listed in Figure 9. From Figure 10, it can be seen that the proportion of intensity differences less than 1 shows a gradual upward trend, with a faster rate of increase in the first 10 s and model 1 showing the most obvious rate of increase. This means that the prediction performance of the three models will have a significant improvement over time and reach over 90% at 20 s, indicating good prediction results.

Figure 10 
                  Proportional changes of predicted intensity differences within 1 for the three models.
Figure 10

Proportional changes of predicted intensity differences within 1 for the three models.

3.3 Overview of vibration table testing and construction of neural networks

Overall, the number of intermediate layers selected for the network used for predicting data is only 1 layer. Selection of input layer nodes: In the vibration table experiment, there are 13 acceleration measurement points, each of which contains 63,877 acceleration values. When the number of samples reaches a certain level, the speed of the network is also difficult to improve [40], and the training error will not decrease. Therefore, the correlation coefficient method is used to select the training set, and the correlation coefficient calculation formula is

(7) r x , y = cov ( x , y ) σ x σ y = 1 n i = 1 n ( x i x ¯ ) ( y i y ¯ ) 1 n i = 1 n ( x i x ¯ ) 2 1 n i = 1 n ( y i y ¯ ) 2 ,

where r x , y represents the correlation coefficient between x and y, I is a positive integer, X is the average value of x, and Y is the average value of y. The correlation coefficient represents not only the direction of linear correlation but also the degree of linear correlation, with a value interval of [−1, 1]. When the correlation coefficient is 1, the two variables are positively correlated, and when the value is −1, the two variables are negatively correlated. The closer the absolute value of the correlation coefficient is to 1, the more obvious the linear relationship is [40]. The empirical formula for the number of hidden layer nodes is

(8) l = n + m + a ,

where l is the number of hidden layer nodes, n is the number of output layer nodes, m is the number of input layer nodes, and a is a constant between [1,10]. After determining the value range of l according to Eq. (8), use the trial and error method to determine the number of hidden layer nodes: first, set a larger number of hidden layer nodes, then gradually reduce the number of nodes and train with the same sample to determine the number of nodes corresponding to the smallest network error. The process of obtaining neural network prediction data is as follows: first, train the neural network by randomly selecting the acceleration measurement point data with complete measurement, determine the topological structure of the neural network by trial and error method, start from the larger network topological structure, gradually reduce the network structure, and obtain the best prediction network structure based on the average absolute error (MAE) and root mean square error (RMSE); Second, another group of fully measured acceleration data validation group is selected to verify the applicability of the network structure. Finally, use a network structure with high prediction accuracy to predict unmeasured data in the experiment. The correlation coefficients between the acceleration data of point 5 in working condition 2 and the data of other measuring points are listed in Table 2.

Table 2

Correlation coefficient between test point 5 and other test points under condition 2

Correlation coefficient Acceleration measuring point Seismic table acceleration
1 2 3 4 5 6 7 8 9 10 11 12
R x,y 0.2029 0.5605 0.9401 0.9921 0.9967 0.9829 0.9870 0.9856 0.9722 0.8842 0.9772 0.2508 0.9005

From Table 2, it can be seen that there is a significant correlation between acceleration measurement point 4, measurement point 6, and seismic table acceleration in condition 2 and measurement point 5, all of which are greater than 0.98. Therefore, measuring point 4, measuring point 6, and seismic table acceleration values were selected as input values for the neural network. The entire process acceleration data were divided into a training set and a testing set in a 3:2 ratio, and the neural network was trained (63,876 sampling points). From equation (8), it can be seen that when the input node is 3, the selection range for the number of hidden layer nodes in the network is [3,12]. During the training process, it was found that when the number of hidden layer nodes is seven, the training frequency is the least and the target accuracy is achieved. Therefore, the hidden layer nodes are determined to be 7. The RMSE and MAE are used to evaluate the prediction accuracy, and their calculation formulas are

(9) RMSE = 1 n i = 1 n ( Y i Y ˆ i ) 2 ,

(10) MAE = 1 n i = 1 n | Y i Y ˆ i | ,

where Y i is the expected output and i is the predicted value.

The acceleration data were predicted, and the RMSE and MAE comparison results between the predicted values and the true values are shown in Table 3.

Table 3

Error comparison between the predicted value and the actual value at test point 5

Prediction method Error term The input layer node is 3 The input layer node is 2
BP neural network MAE 56.6504 1.9560
RMSE 234.4247 9.1665
Our method MAE 36.1566 0.7422
RMSE 158.8076 3.5358

From Table 3, it can be seen that when the number of input layer nodes is 3, the MAE between the predicted acceleration value of the Our method neural network and the true value is 36.1566, which is 63.82% of the MAE between the predicted and true values of the BP neural network. The RMSE is 158.8076, which is 67.78% of the BP neural network and has a significant error. When the number of input layer nodes was reduced to 2 and 1 for network training, it was found that the prediction accuracy of the network was lower when the number of input layer nodes was 1 than when the number of input layer nodes was 3. Therefore, the study only provided prediction accuracy analysis when the number of input layer nodes was 2. Select measurement point 4 and measurement point 6 with the highest correlation coefficient with measurement point 5 as the input set, use the trial and error method to determine the hidden layer as 6, and train the network.

3.4 Topological construction of neural networks

Based on Table 2 and Figure 3, it can be seen that when the number of input nodes is 2, the prediction accuracy of the OURS method neural network is better than that of the three-layer input. In summary, it can be seen that when the number of input nodes is 2, the prediction results of the OURS method neural network topology for acceleration data are basically consistent with the actual values. Therefore, a three-layer BP neural network optimized by MAE is used to predict vibration data, with two input nodes, six hidden layer nodes, and one output layer node. The construction process of a neural network for predicting acceleration data is shown in Figure 11.

Figure 11 
                  Neural network acceleration data prediction process.
Figure 11

Neural network acceleration data prediction process.

After the network training is completed, its applicability needs to be verified. Select the acceleration measurement point 9 under the same working condition, i.e., working condition 2, as the validation group, with 40 data points exceeding ±80 m/s as the predicted data. Use the trained neural network topology structure mentioned above for prediction. The correlation coefficients between measurement point 9 in condition 2 and other measurement points are listed in Table 4.

Table 4

Correlation coefficient between test point 9 and each test point under condition 2

Correlation coefficient Acceleration measuring point Seismic table acceleration
1 2 3 4 5 6 7 8 9 10 11 12
R x,y 0.1804 0.5386 0.9144 0.9700 0.9856 0.9916 0.9870 0.9887 0.9987 0.9933 0.9772 0.8922 0.9005

From Table 4, it can be seen that in condition 2, the correlation coefficients between measurement points 8, 9, and 10 are the highest and all greater than 0.99. Therefore, in the training set, measurement points 8 and 10 are selected as the input sets for the network topology structure obtained above. The residual value between the predicted value and the true value of measurement point 9 is shown in Figure 5. From Figure 12, it can be seen that out of the 40 predicted data points at measurement point 9, the maximum residual between the predicted value and the true value is 1.70817 m/s2, which occurred in the 17.226th second. The MAE between the predicted value and the true value is 0.438, and the RMSE is 3.765, with an acceptable error range. Therefore, the topology structure of the OURS method neural network trained above is applicable for predicting acceleration data and can be used to predict 29 data points beyond the range of the accelerometer in the experiment (Figure 13).

Figure 12 
                  Residual value between the predicted value and the real value of test point 9.
Figure 12

Residual value between the predicted value and the real value of test point 9.

Figure 13 
                  Acceleration time history curve and partially enlarged view of measurement point 7 under condition 2 after completion. (a) Acceleration process curve, (b) A locally enlarged image, (c) B locally enlarged image.
Figure 13

Acceleration time history curve and partially enlarged view of measurement point 7 under condition 2 after completion. (a) Acceleration process curve, (b) A locally enlarged image, (c) B locally enlarged image.

3.5 Predicted results of unmeasured data

There are 15 data values of 90 m/s2 and 14 data values of −90m/s2 in the acceleration response time range of point 7 in working condition 2. Using the previously trained and validated neural network topology, select the neural network training set using the correlation coefficient method to predict measurement points using unmeasured data.

The correlation coefficients between other measurement points in condition 2 and measurement point 7 are listed in Table 5. Select measurement points 6 and 8 with a correlation coefficient greater than 0.99 as the input sets.

Table 5

Correlation coefficient between test point 7 and other test points under condition 2

Correlation coefficient Acceleration measuring point Seismic table acceleration
1 2 3 4 5 6 7 8 9 10 11 12
Rx,y 0.1689 0.5208 0.9088 0.9705 0.9870 0.9916 0.9921 0.9987 0.9822 0.9901 0.9772 0.2207 0.9865

3.6 Visualization denoising of synthetic seismic signals

In order to compare and analyze the denoising effect of algorithms from both qualitative and quantitative perspectives, the trained model is applied to the test set. Add Gaussian white noise with noise levels of 5, 8, and 10% to the simple model data and further evaluate the denoising effect of the proposed method and Unet under different noise levels. The denoising results are shown in Figure 14 when the noise level is 8%. From Figure 14, it can be seen that, compared to Unet, the proposed method removes most of the noise, and the denoised profile is closer to the original clean signal.

Figure 14 
                  Comparison of synthetic seismic signal denoising results (a. clean signal; b. noisy signal; c. Unet denoising result; d. residual profile of OURS method enhancement).
Figure 14

Comparison of synthetic seismic signal denoising results (a. clean signal; b. noisy signal; c. Unet denoising result; d. residual profile of OURS method enhancement).

4 Conclusion

Compared with the traditional CNN, the model proposed in this article has its unique advantages. It can deal with the problem of long-term sequence dependence and can carry out parallel computing. It can focus on global information and has a broader receptive field, making the inversion network a greater advantage in the seismic inversion of long-term sequence. The application of model data testing and actual data also indicates that this method can achieve good inversion results in seismic impedance inversion, providing a new method for intelligent seismic reservoir prediction. This method can effectively enhance weak reflection, improve the recognition ability of seismic data for weak reflection interfaces, and provide a feasible solution for small-layer division with broad application prospects.

  1. Funding information: This work has not received any external funding.

  2. Author contributions: Dingbo Chen conducted experimental comparisons. Liangjun Lai conducted data collection and analysis.

  3. Conflict of interest: The authors declare that there is no conflict of interest.

References

[1] Li W. Rupture characteristics and seismic assessment of the Qinling northern piedmont fault. Nat Hazards Res. 2022;2(3):183–92.10.1016/j.nhres.2022.07.004Search in Google Scholar

[2] Farra V, Psencik I. Weak-anisotropy approximation of P-wave geometrical spreading in horizontally layered anisotropic media of arbitrary symmetry: TTI specification. Geophysics. 2021;86(4):1–61.10.1190/geo2020-0720.1Search in Google Scholar

[3] Li X, Xie M, Zhang Y, Ding G, Tong W. Dual attention convolutional network for action recognition. IET Image Process. 2020;14(6):1059–65.10.1049/iet-ipr.2019.0963Search in Google Scholar

[4] Klaeschen D, Belykh I, Gnibidenko H, Patrikeyev S, von Huene R. Structure of the Kuril Trench from seismic reflection records. J Geophys Res Solid Earth. 1994;99(B12):24173–88.10.1029/94JB01186Search in Google Scholar

[5] Xiaofei G, Tonggeng O, Wugang M, Linbin W, Yifei Z, Jun L, et al. A new method for suppressing random noise in seismic signals based on CEEMDAN and wavelet transform. Geod Geodyn. 2022;42(11):1202–6 + 1210.Search in Google Scholar

[6] Zou Z, Zhang L. Enhancing low-frequency water-column acoustic reflections in marine multichannel seismic data for seismic oceanography. J Acoust Soc Am. 2021;150(5):3852–60.10.1121/10.0007278Search in Google Scholar PubMed

[7] Kunhong L. Hidden feature extraction and pattern analysis of high dimensional seismic signals [dissertation]. Chengdu: University of Electronic Science and Technology, 2022.Search in Google Scholar

[8] Zhang Q, Zhang J, Lei Q, Bloomberg, Liu S. Parallel algorithm for frequency compensation of seismic signals based on CUDA. Pet Geophys Explor. 2022;57(5):1241–9 + 1010.Search in Google Scholar

[9] Li X, Ma F, Deng L, Yang M, Chen X. Low-order approximation of thin-layer reflection and transmission coefficients with respect to ray parameters. Geophys Prospect Pet. 2022;58(5):654–60.Search in Google Scholar

[10] Barreca G, Gross F, Scarfì L, Aloisi M, Monaco C, Krastel S. The Strait of Messina: Seismotectonics and the source of the 1908 earthquake. Earth-Sci Rev. 2021;218(3):103685.10.1016/j.earscirev.2021.103685Search in Google Scholar

[11] Jiang L, Shang W, Zhou J, Wei M, Wang Y. A second-order compression wavelet transform algorithm for seismic signal analysis. J Electron Inf Technol. 2021;43(12):3710–7.Search in Google Scholar

[12] Luo H, Zhao C. Research on filtering performance of seismic signal filters. J Chongqing Univ Sci Technol (Nat Sci Ed). 2021;23(06):76–82.Search in Google Scholar

[13] Qiao Y, Li Q, Qian H, Song X. Research on seismic signal denoising method based on VMD and improved wavelet threshold. Geophys Geochem Explor Comput Technol. 2021;43(06):690–6.Search in Google Scholar

[14] Zhang Q, Ye Z, Shao S, Niu T, Zhao Y. Remaining useful life prediction of rolling bearings based on convolutional recurrent attention network. Assem Autom. 2022;3:42–387.10.1108/AA-08-2021-0113Search in Google Scholar

[15] Margheriti L, Nostro C, Cocina O, Castellano M, Moretti M, Lauciani V, et al. Seismic surveillance and earthquake monitoring in Italy. Seismol Res Lett. 2021;3:92–1671.10.1785/0220200380Search in Google Scholar

[16] Guo MH, Lu CZ, Hou Q, Liu Z, Cheng MM, Hu SM. Semantic: Rethinking convolutional attention design for semantic segmentation. Adv Neural Inf Process Syst. 2022;35:1140–56.Search in Google Scholar

[17] Shi J, Huang H, Xue S, Li B. Research on seismic signal classification and recognition based on PSO-SVM. Comput Digital Eng. 2021;49(6):1037–41 + 1046.Search in Google Scholar

[18] Shi J, Huang H, Xue S, Li B, Yuan X. Research on seismic signal classification based on EMD-VMD-LSTM. Sens Microsyst. 2021;40(6):57–9 + 62.Search in Google Scholar

[19] Wang Y. Application of convolutional neural networks based on low rank decomposition in time and frequency domain in desert seismic signal processing [dissertation]. Changchun: Jilin University. 2021. 10.27162/d.cnki.gjlin.2021.001878.Search in Google Scholar

[20] Zhang S. Research on seismic signal noise reduction algorithms based on deep learning [dissertation]. Sanhe City: School of Disaster Prevention and Technology. 2021.Search in Google Scholar

[21] Tang X, Yang Y, Yang W, Lanning J, Chen YF. Experimental and numerical investigation on the seismic behavior of plane frames with special-shaped concrete-filled steel tubular columns. J Build Eng. 2021;35(1):102070.10.1016/j.jobe.2020.102070Search in Google Scholar

[22] Homrighausen S, Hoernle K, Wartho JA, Hauff F, Werner R. Do the 85°E Ridge and Conrad Rise form a hotspot track crossing the Indian Ocean? Lithos. 2021;398–399:106234.10.1016/j.lithos.2021.106234Search in Google Scholar

[23] Stannard D, Meyers J, Scopel A. The Abra sedimentary-hosted Pb-Ag-Cu-Au deposit, Western Australia: A geophysical case study. Lead Edge. 202140(2):129–38.10.1190/tle40020129.1Search in Google Scholar

[24] Rad B. Improving 3D water column seismic imaging using the Common Reflection Surface method. J Appl Geophys. 2020;179(1):356–63.10.1016/j.jappgeo.2020.104072Search in Google Scholar

[25] Wei R, Sun J, Meng X, Xu Y, Liu M, Shang Y. Analysis of the impact of velocity smoothing on reflected seismic signals. J Geophys. 2020;63(10):3849–56.Search in Google Scholar

[26] Tian Y, Gao J, Wang D. Improving seismic resolution based on enhanced multi-channel variational mode decomposition. J Appl Geophys. 2022;199:104592.10.1016/j.jappgeo.2022.104592Search in Google Scholar

[27] Lu X, Hu X, Zhang J, Liao X, Fu Z. A continuous towed seismic data acquisition and correlation stack method for rapid detection of urban road. J Appl Geophys. 2022;199:199.10.1016/j.jappgeo.2022.104585Search in Google Scholar

[28] Wang Z. Research on seismic signal feature extraction and clustering analysis method based on convolutional neural network [dissertation]. Beijing: China University of Petroleum; 2020.Search in Google Scholar

[29] Shen L. Wave velocity analysis method for tunnel seismic signals based on polarization analysis [dissertation]. Chongqing: Chongqing University; 2020.Search in Google Scholar

[30] Song Q. Research on seismic signal denoising method based on modal decomposition technology [dissertation]. Daqing: Northeast Petroleum University; 2020.Search in Google Scholar

[31] Cao Y, Fu Y, Zhu Z, Rao Z. Color random valued impulse noise removal based on quaternion convolutional attention denoising network. IEEE Signal Process Lett. 2022;29:29–373.10.1109/LSP.2021.3135803Search in Google Scholar

[32] Wang Y, Li H, Tuo X, Shen T. A new method for picking up P-phase first breaks from strong noise microseismic signals. Pet Geophys Prospect. 2020;59(03):356–65.Search in Google Scholar

[33] Xu J. Research on seismic signal analysis of high [dissertation]. Xi’An: Chang’an University; 2020. 10.26976/d.cnki.gchau.2020.001673.Search in Google Scholar

[34] Peng J. Research on feature reconstruction and clustering analysis method of prestack seismic signals based on sparse coding [dissertation]. Beijing: China University of Petroleum; 2020. 10.27643/d.cnki.gsybu.2020.001543.Search in Google Scholar

[35] Takahashi T, Kaiho Y, Ishihara Y, Obana K, Miura S, Kodaira S, et al. Trans-dimensional imaging of the random inhomogeneity structure in the southern Ryukyu arc, Japan. Geophys J Int. 2021;11(07):512–21.Search in Google Scholar

[36] Zhu Y, Zhao C, Guo H, Wang J, Zhao X, Lu H. Attention CoupleNet: Fully convolutional attention coupling network for object detection. IEEE Trans Image Process. 2018;28(1):113–26.10.1109/TIP.2018.2865280Search in Google Scholar PubMed

[37] Wu J, Chen X, Shi H, Hu G, Ye J, Li Y, et al. Identification of pesticide poisoning in silkworm based on convolutional neural network. Silkworm Sci Commun. 2022;42(04):17–23.Search in Google Scholar

[38] Chen K, Li J, Chang Q, Xu W, Pang G. Research progress in rapid detection of pesticide residues based on molecular recognition. J Anal Test. 2022;41(09):1439–46.Search in Google Scholar

[39] Kong R. Identification of high-risk pesticides in the water body of Liangzi Lake and their impact on aquatic ecological risks [dissertation]. Wuhan: Huazhong Agricultural University; 2022. 10.27158/d.cnki.ghznu.2022.000226.Search in Google Scholar

[40] Hajnal Z, White DJ, Takacs E, Gyorfi I, Annesley IR, Wood G, et al. Application of modern 2-D and 3-D seismic-reflection techniques for uranium exploration in the Athabasca Basin. Can J Earth Sci. 2010;47(5):761–82.10.1139/E10-026Search in Google Scholar

Received: 2023-09-07
Accepted: 2023-09-22
Published Online: 2024-01-09

© 2024 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Editorial
  2. Focus on NLENG 2023 Volume 12 Issue 1
  3. Research Articles
  4. Seismic vulnerability signal analysis of low tower cable-stayed bridges method based on convolutional attention network
  5. Robust passivity-based nonlinear controller design for bilateral teleoperation system under variable time delay and variable load disturbance
  6. A physically consistent AI-based SPH emulator for computational fluid dynamics
  7. Asymmetrical novel hyperchaotic system with two exponential functions and an application to image encryption
  8. A novel framework for effective structural vulnerability assessment of tubular structures using machine learning algorithms (GA and ANN) for hybrid simulations
  9. Flow and irreversible mechanism of pure and hybridized non-Newtonian nanofluids through elastic surfaces with melting effects
  10. Stability analysis of the corruption dynamics under fractional-order interventions
  11. Solutions of certain initial-boundary value problems via a new extended Laplace transform
  12. Numerical solution of two-dimensional fractional differential equations using Laplace transform with residual power series method
  13. Fractional-order lead networks to avoid limit cycle in control loops with dead zone and plant servo system
  14. Modeling anomalous transport in fractal porous media: A study of fractional diffusion PDEs using numerical method
  15. Analysis of nonlinear dynamics of RC slabs under blast loads: A hybrid machine learning approach
  16. On theoretical and numerical analysis of fractal--fractional non-linear hybrid differential equations
  17. Traveling wave solutions, numerical solutions, and stability analysis of the (2+1) conformal time-fractional generalized q-deformed sinh-Gordon equation
  18. Influence of damage on large displacement buckling analysis of beams
  19. Approximate numerical procedures for the Navier–Stokes system through the generalized method of lines
  20. Mathematical analysis of a combustible viscoelastic material in a cylindrical channel taking into account induced electric field: A spectral approach
  21. A new operational matrix method to solve nonlinear fractional differential equations
  22. New solutions for the generalized q-deformed wave equation with q-translation symmetry
  23. Optimize the corrosion behaviour and mechanical properties of AISI 316 stainless steel under heat treatment and previous cold working
  24. Soliton dynamics of the KdV–mKdV equation using three distinct exact methods in nonlinear phenomena
  25. Investigation of the lubrication performance of a marine diesel engine crankshaft using a thermo-electrohydrodynamic model
  26. Modeling credit risk with mixed fractional Brownian motion: An application to barrier options
  27. Method of feature extraction of abnormal communication signal in network based on nonlinear technology
  28. An innovative binocular vision-based method for displacement measurement in membrane structures
  29. An analysis of exponential kernel fractional difference operator for delta positivity
  30. Novel analytic solutions of strain wave model in micro-structured solids
  31. Conditions for the existence of soliton solutions: An analysis of coefficients in the generalized Wu–Zhang system and generalized Sawada–Kotera model
  32. Scale-3 Haar wavelet-based method of fractal-fractional differential equations with power law kernel and exponential decay kernel
  33. Non-linear influences of track dynamic irregularities on vertical levelling loss of heavy-haul railway track geometry under cyclic loadings
  34. Fast analysis approach for instability problems of thin shells utilizing ANNs and a Bayesian regularization back-propagation algorithm
  35. Validity and error analysis of calculating matrix exponential function and vector product
  36. Optimizing execution time and cost while scheduling scientific workflow in edge data center with fault tolerance awareness
  37. Estimating the dynamics of the drinking epidemic model with control interventions: A sensitivity analysis
  38. Online and offline physical education quality assessment based on mobile edge computing
  39. Discovering optical solutions to a nonlinear Schrödinger equation and its bifurcation and chaos analysis
  40. New convolved Fibonacci collocation procedure for the Fitzhugh–Nagumo non-linear equation
  41. Study of weakly nonlinear double-diffusive magneto-convection with throughflow under concentration modulation
  42. Variable sampling time discrete sliding mode control for a flapping wing micro air vehicle using flapping frequency as the control input
  43. Error analysis of arbitrarily high-order stepping schemes for fractional integro-differential equations with weakly singular kernels
  44. Solitary and periodic pattern solutions for time-fractional generalized nonlinear Schrödinger equation
  45. An unconditionally stable numerical scheme for solving nonlinear Fisher equation
  46. Effect of modulated boundary on heat and mass transport of Walter-B viscoelastic fluid saturated in porous medium
  47. Analysis of heat mass transfer in a squeezed Carreau nanofluid flow due to a sensor surface with variable thermal conductivity
  48. Navigating waves: Advancing ocean dynamics through the nonlinear Schrödinger equation
  49. Experimental and numerical investigations into torsional-flexural behaviours of railway composite sleepers and bearers
  50. Novel dynamics of the fractional KFG equation through the unified and unified solver schemes with stability and multistability analysis
  51. Analysis of the magnetohydrodynamic effects on non-Newtonian fluid flow in an inclined non-uniform channel under long-wavelength, low-Reynolds number conditions
  52. Convergence analysis of non-matching finite elements for a linear monotone additive Schwarz scheme for semi-linear elliptic problems
  53. Global well-posedness and exponential decay estimates for semilinear Newell–Whitehead–Segel equation
  54. Petrov-Galerkin method for small deflections in fourth-order beam equations in civil engineering
  55. Solution of third-order nonlinear integro-differential equations with parallel computing for intelligent IoT and wireless networks using the Haar wavelet method
  56. Mathematical modeling and computational analysis of hepatitis B virus transmission using the higher-order Galerkin scheme
  57. Mathematical model based on nonlinear differential equations and its control algorithm
  58. Bifurcation and chaos: Unraveling soliton solutions in a couple fractional-order nonlinear evolution equation
  59. Space–time variable-order carbon nanotube model using modified Atangana–Baleanu–Caputo derivative
  60. Minimal universal laser network model: Synchronization, extreme events, and multistability
  61. Valuation of forward start option with mean reverting stock model for uncertain markets
  62. Geometric nonlinear analysis based on the generalized displacement control method and orthogonal iteration
  63. Fuzzy neural network with backpropagation for fuzzy quadratic programming problems and portfolio optimization problems
  64. B-spline curve theory: An overview and applications in real life
  65. Nonlinearity modeling for online estimation of industrial cooling fan speed subject to model uncertainties and state-dependent measurement noise
  66. Quantitative analysis and modeling of ride sharing behavior based on internet of vehicles
  67. Review Article
  68. Bond performance of recycled coarse aggregate concrete with rebar under freeze–thaw environment: A review
  69. Retraction
  70. Retraction of “Convolutional neural network for UAV image processing and navigation in tree plantations based on deep learning”
  71. Special Issue: Dynamic Engineering and Control Methods for the Nonlinear Systems - Part II
  72. Improved nonlinear model predictive control with inequality constraints using particle filtering for nonlinear and highly coupled dynamical systems
  73. Anti-control of Hopf bifurcation for a chaotic system
  74. Special Issue: Decision and Control in Nonlinear Systems - Part I
  75. Addressing target loss and actuator saturation in visual servoing of multirotors: A nonrecursive augmented dynamics control approach
  76. Collaborative control of multi-manipulator systems in intelligent manufacturing based on event-triggered and adaptive strategy
  77. Greenhouse monitoring system integrating NB-IOT technology and a cloud service framework
  78. Special Issue: Unleashing the Power of AI and ML in Dynamical System Research
  79. Computational analysis of the Covid-19 model using the continuous Galerkin–Petrov scheme
  80. Special Issue: Nonlinear Analysis and Design of Communication Networks for IoT Applications - Part I
  81. Research on the role of multi-sensor system information fusion in improving hardware control accuracy of intelligent system
  82. Advanced integration of IoT and AI algorithms for comprehensive smart meter data analysis in smart grids
Downloaded on 8.12.2025 from https://www.degruyterbrill.com/document/doi/10.1515/nleng-2022-0336/html
Scroll to top button