Startseite 3D reconstruction study of motion blur non-coded targets based on the iterative relaxation method
Artikel Open Access

3D reconstruction study of motion blur non-coded targets based on the iterative relaxation method

  • Shi Yun , Chen Rongna und Zhu Yanyan EMAIL logo
Veröffentlicht/Copyright: 30. August 2025
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

Achieving high-quality three-dimensional (3D) reconstruction has been a challenging problem due to factors such as motion blur. In this article, we first construct a mathematical model of an iterative relaxation method in reconstructing images, including iterative method and relaxation method. Then, the iterative image derivation model with relaxation factors is constructed by introducing relaxation factors. Next, a motion blur model based on the iterative relaxation method is proposed and combined with the 3D reconstruction method of non-coding points. Finally, the 3D reconstruction in some real scenes is carried out using the motion blur non-coded target 3D reconstruction method based on the iterative relaxation method with error analysis. The results show that the reconstruction accuracy under the optimized path has been improved by two orders of magnitude compared with that under the initial path, and the reconstruction error is basically maintained at about 0.042 mm, with the maximum not exceeding 0.05 mm. This indicates that the proposed method can effectively reduce the reconstruction error and achieve a high reconstruction accuracy. The motion blur non-coded target 3D reconstruction method based on the iterative relaxation method proposed in this article has certain practicality and promotion value.

1 Introduction

With the continuous progress of technology, three-dimensional (3D) reconstruction has been used in medical imaging, virtual reality, and other fields [1,2]. However, in practical applications, the motion of the target leads to motion blur in the image, which affects the accuracy and precision of 3D reconstruction [3]. For static targets of conventional size, the accuracy of 3D shape reconstruction by multi-view images can reach the order of 0.01 mm when supplemented with structured light projection textures or with active visual features laid on the object surface [4,5]. Motion as an important feature of active targets in nature, the application demand of 3D shape measurement of moving targets has gradually promoted the research investment in this field [6,7]. Therefore, how to address the effect of motion blur on 3D reconstruction has become a popular research area.

3D reconstruction technique is one of the research hotspots in the field of 3D profile measurement. Liu et al. proposed a framework for uncertainty assessment based on the Bayesian approach by investigating various uncertainties in the 3D reconstruction process, such as pixel noise and symmetry mismatch [8]. Nicolosi and Spena used high-resolution magnetic resonance imaging and computer-aided design software to reconstruct the patient’s neuroanatomical structure into a 3D model and embedding it into a surgical navigation system [9]. Xie reconstructed a 3D model of the target object by processing multiple viewpoint images during cinematography using photometric stereo fusion [10]. Tarsitano et al. reconstructed a 3D model of a patient’s CT image by converting it into a 3D model and using the CAD/CAM (Computer Aided Design/Computer Aided Manufacturing) technique for bone reconstruction and then comparing and evaluating the reconstructed results with the actual patient’s condition [11].

Since exposure time is positively correlated with image signal-to-noise ratio, photographing high-speed motion targets by shortening exposure time can severely degrade the signal-to-noise ratio of images [12]. Ali and Mahmood derived the advantages, disadvantages, and applicability of each operator by conducting experiments on a large data set and proposed a motion blur image segmentation method based on multiple blur metric operators [13]. Kubota et al. achieved real-time monitoring and compensation of motion blur of high-speed moving targets by using a combination of a vibrating mirror and a thermal imaging camera [14]. Zhang et al. will automatically separate different images before and after the blurred motion and then use convolutional neural network (CNN) and recurrent neural networks (RNN) for feature extraction and sequence modeling of different images to finally obtain high-quality 3D reconstruction results [15].

In this article, first, relaxation factors are introduced to construct an iterative image derivation model with relaxation factors, and then a motion blur model based on the iterative relaxation method is proposed. Second, the non-coded point 3D reconstruction method is combined with the motion blur model based on iterative relaxation method to realize the 3D reconstruction.

2 Mathematical model of the iterative relaxation method in reconstructing images

2.1 Iterative relaxation method construction

2.1.1 Iterative method

The iterative method starts from a given initial vector x ( 0 ) and constructs an appropriate iterative formula that gradually computes the vector x ( 1 ) , x ( 2 ) , such that the sequence of vectors { x ( k ) } converges to an exact solution. In this way, for a suitably large k , x ( k ) can be taken as an approximate solution. After the initial vector x ( 0 ) is determined, the estimated value vector can be derived by the following equation:

(1) T x ( k ) = T x ( k 1 ) + c ,

where k = 1 , 2 , 3 , . The surface can be assumed to be a cartesian surface z = f ( x , y ) , and the normal to the surface can be replaced by n ( p , q , 1 ) , where p = H x , q = H y , establishing the origin coordinates as shown in Figure 1, where n represents the direction of the normal, θ is the angle of the normal to the z axis, and ϕ is the angle of the normal plane to the x axis. Assuming that the initial vector u ( 0 ) D is given, the following iterative algorithm is used to calculate u ( k + 1 ) when k = 0 , 1 , 2 , until u ( k ) the termination condition is satisfied:

(2) W x ( k + 1 ) = ( 1 α ) W x ( k ) + α T y ( k ) + α P ( u ( k ) ) W y ( k + 1 ) = β T x ( k ) + ( 1 β ) W y ( k ) + α Q ( u ( k ) )

where the constant α > 0 , β > 0 .

Figure 1 
                     The coordinate of surface material.
Figure 1

The coordinate of surface material.

2.1.2 Relaxation method

The relaxation method is an accelerated iterative method obtained by slightly improving the Gauss–Seidel iteration method, which is one of the effective methods for solving large systems of sparse matrix equations. The result of the k + 1 st iteration obtained by the Gauss–Seidel iteration formula is

(3) x ˆ ( k + 1 ) = b i j = 1 i 1 a i i x j ( k + 1 ) j = i + 1 n a i i x ˆ ( k ) a i i ,

where x ˆ ( k + 1 ) is used as an intermediate value only. The real parameter w is introduced and the weighted average of the result x i ( k + 1 ) of the k th iteration is taken as x ˆ ( k + 1 ) , i.e.

(4) x i ( k + 1 ) = x i ( k ) + w ( x ˆ i ( k + 1 ) x ˆ i ( k ) ) .

The method of calculating the approximate sequence of solutions to a system of n -element linear equations A x = b by the above equation is called the relaxation method, and w is called the relaxation factor. When w < 1 is called the Gauss-Seidel iteration, and w > 1 is called the super-relaxation method, or SOR for short.

2.2 Iterative image export with relaxation factor

From the above relaxation factor construction, it is clear that the projection matrix R is an M × N matrix. Now, multiply both sides of equation (4) by R T at the same time to obtain the homogeneous set of equations of equation (4) as

(5) R T R x = R T y .

Abbreviate equation (5) as

(6) A x = b ,

where A = R T R , b = R T y . Satisfying condition a i j 0 decomposes A into

(7) A = D w + L + 1 1 w D + U ,

where the real constant, w > 0 , is called the relaxation factor. Here, it can be expressed as

(8) D = a 11 a 22 0 a n n L = 0 a 21 0 M 0 0 a n 1 L a n , n 1 0 U = 0 a 12 L a 1 n 0 0 M 0 a n 1 , n 0 . .

From the iterative formula for the system of linear equations, we have

(9) x ( k + 1 ) = L + D w 1 U + 1 1 w ( x ) ( k ) + b L + D w 1 .

The iterative method expressed by iterative equation (9) is called the successive super-relaxation method in order to avoid finding the inverse matrix of L + D w . By making the following change to its iterative equation (9) with L + D w left and collapsing it, we obtain

(10) x i ( k + 1 ) = x i ( k ) + w s = 1 M r s i ( y s j = i N x j ( k ) r s i j = i i 1 x j ( k + 1 ) r s i ) s = 1 M r s i 2 .

Due to the special nature of the projection matrix storage, the formula is already very large for just 128 × 128 pictures, which not only takes tens of hours to compute, but also requires a lot of memory space. In practical applications, the size of images is usually 512 × 512 or larger. Therefore, it is unrealistic to use this equation in image reconstruction, and for this reason, equation (10) is modified as follows:

(11) x i ( k + 1 ) = x i ( k ) + w s = 1 M r s i 2 s = 1 M r s i ( y s R 2 x ( k ) ) .

The algorithm derived from equations (10) and (11) is called the iterative method with relaxation factor. By changing x j ( k + 1 ) to x j ( k ) in equation (10), the last two terms in equation (11) are combined into R s x ( k ) . This allows the ( k + 1 ) th iteration to calculate the approximation of the image vector at the same time when the program is executed, instead of calculating x j ( k + 1 ) 1 and then x j ( k + 1 ) , thus greatly reducing the reconstruction time. The iterative method with relaxation factor is a joint iterative reconstruction method or simultaneous iterative reconstruction method, which is helpful to overcome the sensitivity of the reconstructed image to measurement noise.

3 3D reconstruction method of a motion blur non-coding target based on the iterative relaxation method

3.1 Traditional projective motion path (PMP) motion blur model

The traditional PMP motion blur model holds on the premise that the degradation process of the image is linearly translational invariant; however, this ideal situation is difficult to encounter in real life. This requires that the motion blur image must be partitioned into multiple subregions, and the PMP is consistent at any position in the same subregion. However, such a process makes the whole modeling process tedious and computationally intensive.

Assume that the pixel intensity of any pixel point on an image is determined by the intensity of light received by the photosensitive element on the imaging sensor during the exposure time. Then, the imaging process can also be expressed as

(12) I ( x ) = 0 T σ I ( x , t ) d t i = 1 N I ( x , t i ) ,

where I ( x ) is the captured image, σ I ( x , t ) d t is the image recorded by the camera during an infinitesimal time interval d t starting from a certain moment t . x is the zeta image coordinate of 3 × 1 and T represents the exposure time. If N moments are sampled within the exposure time and N is large enough, then σ I ( x , t ) d t can be I ( x , t i ) approximated instead, so that the integral operation of the entire image can be expressed as a summation approximation.

Disregarding noise, when there is no relative motion between the camera and the scene being photographed, there is

(13) I ( x , t 1 ) I ( x , t 2 ) I ( x , t N ) .

When there is relative motion between the camera and the scene being captured, I ( x ) is equal to the sum of I ( x , t i ) for multiple unaligned images. The PMP motion blur model uses I ( x , t i ) = I ( h i x , t i ) to represent the relative position relationship between the unaligned image at the current moment and the corresponding pixels of the image at the previous moment, where h i is a 3 × 3 single response matrix. Assuming that all h i are known, the clear imaging I ( x , t i ) at any moment can be I 0 ( x ) expressed as

(14) I ( x , t i ) = I 0 ( H i x ) N .

Therefore, the PMP motion blur model can be expressed as

(15) B ( y ) = i = 1 N I ( x , t i ) = i = 1 N I 0 ( H i x ) N ,

where B ( y ) is the motion blur image, the model represents the motion blur image as a superimposed average of multiple clear images, each of which has a unique planar projection transformation relationship with I 0 ( x ) .

3.2 Motion blur model based on the iterative relaxation method

Unlike the traditional PMP model, our model based on the iterative relaxation method effectively restores the motion of the captured target and the imaging process during the exposure time. Our model based on the iterative relaxation method is shown in Figure 2, because the motion blur effect of the image is formed by the instantaneous exposure integral of the target along the motion path during the exposure time. Therefore, the exposure time can be discretely sampled into N moments, N large enough. The “reference position” of the selected non-coded point is the center of the circle with the origin O w of the world coordinate system O w X w Y w Z w and the plane of the non-coded point is located in the plane X w O w Y w . At moment i ( i = 1 , 2 , N ) , the non-coded point is moved from the reference position to the current “moment” position by a rotation transformation R i and a translation transformation t i . This is then projected onto the camera imaging plane through a perspective transformation to obtain a clear image of the non-coded point at i . After N different moments in the exposure time, a total of N images are projected on the camera imaging plane, and the average image produced by their superposition is the motion blur image produced during the entire exposure time.

Figure 2 
                  Motion blur model based on the iterative relaxation method.
Figure 2

Motion blur model based on the iterative relaxation method.

According to the camera imaging model, the projection process of the non-coded pattern at moment i can be quantitatively represented as

(16) u ( I i ) v ( L f ) = proj t i + A , k 1 , k 2 , p 1 , p 2 , R , t , R i X Y 0 ,

(17) I ( u ( L f ) , v ( L j ) ) = C ( X , Y , 0 ) ,

where ( X , Y , 0 ) denotes the coordinates of any point on the encoded pattern in the world coordinate system at the reference reference pose. I i is the clear image at the i th sampling moment, ( u ( I i ) , v ( I i ) ) denotes the pixel coordinates in the clear image I i . I ( ( u ( I i ) , v ( I i ) ) ) denotes the gray value of the clear image I i at pixel location ( u ( I i ) , v ( I i ) ) , and C ( X , Y , 0 ) denotes the gray value of the coded pattern at point ( X , Y , 0 ) .

Since the spatial attitude of the encoded point described by the attitude matrix R i can be represented by three angles of rotation around axes X ( w ) , Y ( w ) , and Z ( w ) in turn. The angle of rotation about axis X ( w ) is called the deflection angle, the angle of rotation about axis Y ( w ) is called the pitch angle, and the angle of rotation about Z ( w ) is called the slew angle. In this article, these three angles are written in vector form as r i = r x r y r z T . The spatial position at the current moment i is represented by D ( i ) as the coordinates of that spatial position. For the same moment i , the rotational R i and translational transformations t i are the same for any point on the coded pattern. Therefore, the projection process of all points on the non-coded pattern can be summarized by the function g ( ) , with

(18) I i = g ( A , k 1 , k 2 , p 1 , p 2 , R , t , r i , D i ) .

According to the mechanism of motion blur effect formation, there are

(19) B = i = 1 N I i N ,

where B is the final output blurred image. Substituting equation (18) into equation (19), we obtain

(20) B = i = 1 N g ( A , k 1 , k 2 , p 1 , p 2 , R , t , r i , D i ) N .

The SMP motion blur model established in this article, is related to the pose vector r ( i ) and D ( i ) the spatial position of the target at each moment, in addition to the parameters of the camera itself. r ( i ) describes the motion pose of the target, and D ( i ) describes the motion path of the target.

It can be seen that the motion blur model based on the iterative relaxation method more intuitively expresses the quantitative relationship between the spatial motion path of the target and the motion blur effect.

3.3 Non-coding point 3D reconstruction

The result of optimizing the initial value of the 3D spatial coordinates of the coded point localization center D cinit is recorded as D copt , and then the relationship between the left and right camera positions in the combined camera imaging model and binocular stereo vision. The pixel coordinates D coptl and D coptr of the optimized localization center D copt of this coding point in the corresponding left and right images, respectively, can be expressed as

(21) D coptl = project ( A c l , k c 1 l , k c 2 l , p c 1 l , p c 2 l , R l , t l , D copt ) D coptr = project ( A c l , k c 1 l , k c 2 l , p c 1 l , p c 2 l , R r l 1 R l , R r l 1 ( t l t r l ) , D copt ) .

For camera calibration, a single response relationship is established between the camera imaging plane and the calibration plate plane. The single-response matrix H is obtained by the correspondence between the two groups of five feature points. Then, the single-response relationship between the pixel coordinate systems O 2 l u l v l and O 2 r u r v r of the left and right corresponding images and the divided area can be established, respectively,

(22) D l = H l D f , D r = H r D f ,

where D f is the spatial coordinate of any point on the delimited area, D l and D r are the pixel coordinates of the imaging point of point D f in the left and right corresponding images, respectively. H l and H r represent the single-response relationship between D f and D l and D r , respectively, since the single-response matrix is invertible.

Therefore, the single-response relationship H l r between the corresponding points D l and D r in the left and right corresponding images can be expressed as

(23) D r = H r H l 1 D l = H l r D l .

When there is occlusion or large image noise in the captured image, it will largely reduce the gray value and the reconstruction drift will occur. At this time, the non-coding point 3D reconstruction process is shown in Figure 3.

Figure 3 
                  Non-coding point 3D reconstruction process.
Figure 3

Non-coding point 3D reconstruction process.

4 Experimental results and analysis of motion target 3D reconstruction simulation

To prove the motion blur non-coded target 3D reconstruction algorithm based on the iterative relaxation, an array of synchronous temporal images of motion targets with different exposure times was actually captured. The intra-frame motion paths of motion blur non-coded points laid on the motion targets were reconstructed using the algorithm in this article, and the specific experimental procedure and experimental results are reported in this section.

4.1 Experimental setup

4.1.1 Experimental equipment

A high-speed camera and a laser projection system were used for this experiment. The high-speed camera model is Phantom v2080, with a resolution of 1,080 × 1,960 and a maximum frame rate of 2,50,000 frames/s. The laser projection system uses a DPSS laser with a wavelength of 532 nm and a power of 50 mW.

4.1.2 Experimental sample

A motion target was used for this experiment, on which motion blur non-coded points were laid out. An array of synchronous timing images was taken at different exposure times, with a total of 100 sets of data.

4.1.3 Experimental procedure

Set the camera frame rate, exposure time, gain, and other parameters to obtain clear and stable images. Set the laser wavelength, power, scanning speed, and other parameters to obtain clear and stable motion blur coding points. Lay the motion blur non-coded points on the moving target surface to simulate the actual application scene.

The experimental flow of 3D reconstruction of moving objects in this article is shown in Figure 4, which is generally divided into two parts circled by the dashed boxes (serial numbers 1 and 2, respectively) in the figure. The first part is to calculate the central initial and optimized intra-frame motion paths of motion blur non-coding points using the motion blur non-coding target 3D reconstruction algorithm based on the iterative relaxation method.

Figure 4 
                     Experimental flow of 3D reconstruction of motion target.
Figure 4

Experimental flow of 3D reconstruction of motion target.

4.2 Data acquisition

First, ten simultaneous timing images were acquired at different exposure times to obtain different degrees of motion blur non-coded points. Second, ensure the uniform position and angle of the camera and laser projection system to obtain stable and accurate motion blur non-coded points. Finally, the acquired image data are saved to the computer for subsequent processing.

4.3 Reconstruction results and error analysis

The actual size of all non-coded points used in this article’s experiments are 25 × 25 red spheres at the end of camera calibration. First, the 3D reconstruction technique of motion blur visual features is established to study the 3D spatial coordinate points of these non-coded points at rest. The motion paths of the red blob under the original coordinate axes are shown in Figure 5, where Figure 5(a) shows motion path 1 of the blob at exposure time t , 0.5, 0, 0.5  m in the x axis, respectively. 0.5, 0, 0.5  m in the y axis, respectively. 0, 2, 4, 6, 8 m in the z axis, respectively. The blob was in a spiral pose in the 3D coordinate space, and the camera was calibrated at the end of under the camera calibration, the trajectory of the ball in three rotations was taken. Figure 5(b) shows the motion paths of the red ball in the 3D coordinates after z , with the shooting time interval t being 0.5 s. Finally, the motion paths of each coded point at moment t = 0 for different exposure times were obtained using an iterative relaxation method based on the motion blur non-coding target 3D reconstruction algorithm, including the initial intra-frame motion path (initial path) and the optimized intra-frame motion path (optimized path).

Figure 5 
                  The motion path of the ball under the original coordinate axis.
Figure 5

The motion path of the ball under the original coordinate axis.

Now that the intra-frame motion paths of the non-coded points have been optimized, the blurring degree of the coding point images can be evaluated according to the blurring degree formula defined in the previous section in order to have a more intuitive understanding of the degree of motion blur generated at the exposure times set in this experiment. In this study, the velocity of the moving target can be simulated by extending the exposure time, but considering only the velocity factor is not comprehensive enough. Because speed and exposure time are interrelated, we need to consider the combined index of the two – the blur level. Experiments at different blur levels can more comprehensively verify the adaptability and stability of the algorithm for different situations. Therefore, the degree of fuzziness is more reflective of the situation in real applications and more accurate in assessing the performance and reliability of the algorithm.

The average blur degree [16] of all non-coded targets in the experiment at each exposure time calculated separately and the average reconstruction error of the motion path within the frame calculated are shown in Table 1. Where ε i and ε j denote the average reconstruction errors of the initial path and the optimized path, respectively. It can be seen that the reconstruction accuracy under the optimized path has improved by two orders of magnitude compared with that under the initial path at different exposure times. However, the reconstruction error basically remains around 0.042, and the maximum does not exceed 0.05. The model algorithm execution of the article takes approximately 3 min.

Table 1

Blur degree and motion path reconstruction error at different exposure times

Δ t (ms) 0.5 0 0.2 0.4 0.6
Average blur degree 0.15 0.21 0.34 0.41 0.46
ε i 0.524 0.649 0.513 0.571 0.552
ε j 0.038 0.049 0.035 0.051 0.043

5 Conclusion

A new motion blur non-coding target 3D reconstruction method based on the iterative relaxation method is proposed to improve the traditional PMP motion blur model. The non-coded target 3D reconstruction is combined with the motion blur model based on the iterative relaxation to achieve the 3D reconstruction. Finally, through the analysis of experimental results and errors, the following conclusions are drawn:

  1. In the simulation experiments, the reconstruction accuracies under the optimized paths are all improved by two orders of magnitude compared with those under the initial paths. The reconstruction error is basically maintained around 0.042 mm, with the maximum not exceeding 0.05 mm.

  2. The motion blur model proposed in this article is an ideal state of motion, but in practice, the motion pattern of the target may be more complex. The reconstruction accuracy can be improved by exploring a more accurate relaxation factor selection method.

  1. Funding information: This research was supported by the School-level Research Projects of West Anhui University (WXZR202211), West Anhui University High-level Personnel Research Funding Project (WGKQ2022013, WGKQ2022015), Anhui Provincial Quality Engineering Project (2021sysxzx031, 2022sx171), School-level Quality Engineering Project of West Anhui University (wxxy2022085), the Open Fund of Anhui Undergrowth Crop Intelligent Equipment Engineering Research Center (AUCIEERC-2022-05), University Key Research Project of Department of Education Anhui Province (2024AH051991, 2023AH010078, 2022AH051683), Anhui Province Youth Teacher Training Action - Domestic Visiting Study and Training Funding Project for Young Backbone Teachers (JNFX2023047).

  2. Author contributions: Shi Yun wrote the initial draft, Chen Rongna participated in the project design, and Zhu Yanyan provided research guidance.

  3. Conflict of interest: There are no potential competing interests in this study. All authors have seen the manuscript and approved to submit to your journal. We confirm that the content of the manuscript has not been published or submitted for publication elsewhere.

  4. Data availability statement: Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.

References

[1] F. Porpiglia, E. Checcucci, D. Amparore, M. Manfredi, F. Massa, P. Piazzolla, et al., “Three-dimensional elastic augmented-reality robot-assisted radical prostatectomy using hyperaccuracy three-dimensional reconstruction technology: a step further in the identification of capsular involvement,” J. Eur. Urol., vol. 76, no. 4, pp. 505–514, 2019. 10.1016/j.eururo.2019.03.037Suche in Google Scholar PubMed

[2] M. Mobayen, A. Feizkhah, M. J. Ghazanfari, P. B. Toolaroud, M. Mobayen, J. Osuji, et al., “Intraoperative three-dimensional bioprinting: a transformative technology for burn wound reconstruction,” J. Int. Soc. Burn Injuries, vol. 48, no. 4, pp. 1023–1024, 2022. 10.1016/j.burns.2022.03.014Suche in Google Scholar PubMed

[3] J. Huo and X. Yu, “Three-dimensional mechanical parts reconstruction technology based on two-dimensional image,” Int. J. Adv. Robot. Syst., vol. 17, no. 2, pp. 36–46, 2020. 10.1177/1729881420910008Suche in Google Scholar

[4] B. Oliveira, G.C. Marcellino, P. Rosa, and T. L. F. C. Pinto, “Metrological analysis of the three-dimensional reconstruction based on close-range photogrammetry and the fusion of long-wave infrared and visible-light images,” J. Meas. Sci. Tech., vol. 32, no. 3, (p. 035015 (13 pp), 2021. 10.1088/1361-6501/abb273Suche in Google Scholar

[5] C. T. Seo, S. W. Kang, and M. Cho, “Three-dimensional free view reconstruction in axially distributed image sensing,” J. Chin. Opt. Lett., vol. 8, pp. 40–43, 2017. 10.3788/COL201715.081102Suche in Google Scholar

[6] W.C. Zhao, B. Zhang, C. L. Xu, L. B. Duan, S. M. Wang, “Optical sectioning tomographic reconstruction of three-dimensional flame temperature distribution using single light field camera,” J. IEEE Sens. J. vol. 18, no. 2, pp. 528–539, 2018. 10.1109/JSEN.2017.2772899Suche in Google Scholar

[7] K. Ishikawa, K. Yatabe, and Y. Oikawa, “Physical-model-based reconstruction of axisymmetric three-dimensional sound field from optical interferometric measurement,” J. Meas. Sci. Tech., vol. 32, no. 4, p. 045202 (8 pp), 2021. 10.1088/1361-6501/abce73Suche in Google Scholar

[8] J. Liu, S. Engblom, and C. Nettelblad, “Assessing uncertainties in x-ray single-particle three-dimensional reconstruction,” J. Phys. Rev. E, vol. 98, no. 1, 2018, id. 013303. 10.1103/PhysRevE.98.013303. Suche in Google Scholar PubMed

[9] F. Nicolosi and G. Spena, “Three-dimensional virtual intraoperative reconstruction (3d vir): a novel method to explore a virtual neurosurgical field,” J. World Neurosurg., vol. 137, no. 5, pp. 189–193, 2020. 10.1016/j.wneu.2020.01.112Suche in Google Scholar PubMed

[10] X. L. Xie, “Three-dimensional reconstruction based on multi-view photometric stereo fusion technology in movies special-effect,” J. Multimed. Tools Appl., vol. 79, no. 13–14, pp. 9565–9578, 2020. 10.1007/s11042-019-08034-wSuche in Google Scholar

[11] A. Tarsitano, S. Battaglia, F. Ricotta, B. Bortolani, L. Cercenelli, E. Marcelli, et al., “Accuracy of CAD/CAM mandibular reconstruction: a three-dimensional, fully virtual outcome evaluation method,” J Cranio-Maxillofac. Surg., vol. 46, no. 7, pp. 1121–1125, 2018. 10.1016/j.jcms.2018.05.010Suche in Google Scholar PubMed

[12] M. Suin, K. Purohit, and A. N. Rajagopalan, “Spatially-attentive patch-hierarchical network for adaptive motion deblurring,” In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (14–19 June 2020, VIRTUAL), 2020, pp. 3606–3615. 10.1109/CVPR42600.2020.00366Suche in Google Scholar

[13] U. Ali and M. T. Mahmood, “Analysis of blur measure operators for single image blur segmentation,” J. Appl. Sci., vol. 8, no. 5, p. 807, 2018. 10.3390/app8050807. Suche in Google Scholar

[14] Y. Kubota, T. Hayakawa, Y. Ke, Y. Moko, and M. Ishikawa, “High-speed motion blur compensation system in infrared region using galvanometer mirror and thermography camera,” In: Sensors and Smart Structures Technologies for Civil, Mechanical, and Aerospace Systems, vol. 11379, pp. 154–160, 2020. SPIE. 10.1117/12.2558450Suche in Google Scholar

[15] J. Zhang, K. Yu, Z. Wen, X. Qi, and A. K. Paul, “3D reconstruction for motion blurred images using deep learning-based intelligent systems,” J. Comput. Mater. Contin., vol. 66, no. 2, pp. 2087–2104, 2021. 10.32604/cmc.2020.014220Suche in Google Scholar

[16] Y. Shi and L. Y. Zhang, “Design of Chinese character coded targets for feature point recognition under motion-blur effect,” J. IEEE Access, vol. 8, pp. 124467–124475, 2020. 10.1109/ACCESS.2020.3006020Suche in Google Scholar

Received: 2023-05-25
Revised: 2024-08-14
Accepted: 2024-12-12
Published Online: 2025-08-30

© 2025 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Artikel in diesem Heft

  1. Research Articles
  2. Intelligent data collection algorithm research for WSNs
  3. A novel behavioral health care dataset creation from multiple drug review datasets and drugs prescription using EDA
  4. Speech emotion recognition using long-term average spectrum
  5. PLASMA-Privacy-Preserved Lightweight and Secure Multi-level Authentication scheme for IoMT-based smart healthcare
  6. Blended teaching design of UMU interactive learning platform for cultivating students’ cultural literacy
  7. Basketball action recognition by fusing video recognition techniques with an SSD target detection algorithm
  8. Evaluating impact of different factors on electric vehicle charging demand
  9. An in-depth exploration of supervised and semi-supervised learning on face recognition
  10. The reform of the teaching mode of aesthetic education for university students based on digital media technology
  11. QCI-WSC: Estimation and prediction of QoS confidence interval for web service composition based on Bootstrap
  12. Line segment using displacement prior
  13. 3D reconstruction study of motion blur non-coded targets based on the iterative relaxation method
  14. Overcoming the cold-start challenge in recommender systems: A novel two-stage framework
  15. Optimization of multi-objective recognition based on video tracking technology
  16. An ADMM-based heuristic algorithm for optimization problems over nonconvex second-order cone
  17. A multiscale and dual-loss network for pulmonary nodule classification
  18. Artificial intelligence enabled microgrid power generation prediction
  19. Special Issue on Informatics 2024
  20. Analysis of different IDS-based machine learning models for secure data transmission in IoT networks
  21. Using artificial intelligence tools for level of service classifications within the smart city concept
  22. Applying metaheuristic methods for staffing in railway depots
  23. Interacting with vector databases by means of domain-specific language
  24. Data analysis for efficient dynamic IoT task scheduling in a simulated edge cloud environment
  25. Analysis of the resilience of open source smart home platforms to DDoS attacks
  26. Comparison of various in-order iterator implementations in C++
Heruntergeladen am 27.10.2025 von https://www.degruyterbrill.com/document/doi/10.1515/comp-2024-0020/html
Button zum nach oben scrollen