Home Speech recognition and intelligent translation under multimodal human–computer interaction system
Article Open Access

Speech recognition and intelligent translation under multimodal human–computer interaction system

  • Danhua Huang EMAIL logo and Shuaiqiu Xiang
Published/Copyright: September 3, 2024
Become an author with De Gruyter Brill

Abstract

The traditional translation robot is limited to the translation of single-mode text images and text videos, which has the problem of low translation accuracy. Therefore, speech recognition and intelligent translation in multimodal human–computer interaction (HCI) system are proposed. First, the network structure of speech recognition model in multi-channel HCI system is established, and the multi-head self-attention mechanism is constructed. Then, the artificial intelligence voice wake-up function is designed, and a multimodal machine translation model is constructed. On this basis, selective attention is added to obtain visual recognition of perceived text, and the decoder is used for multimodal gating fusion to realize the output of encoder translation results. Experimental results show that this method has high BLUE value and high translation accuracy.

1 Introduction

Multimodal human–computer interaction (HCI) technology from the character stage, graphical interface stage to the touch screen stage, the touch screen brings convenience while there are also many inconveniences. With the normalization of the Internet and the maturity of artificial intelligence (AI) technology, people’s needs are changing, and they start to explore an interaction mode that can free their hands while also having a more realistic interaction experience [1]. Multimodal HCI is the development trend of natural HCI, and multimodal information interaction is used as input and output to make full use of the characteristics of various modalities and improve the efficiency and realism of HCI. In recent years, the rapid development of machine learning has prompted more and more researchers to consider how to use the increasingly mature computer vision, speech recognition, speech synthesis, and other technologies in the field of natural language processing to realize the application of multimodal HCI [2]. The current continuous improvement in machine learning methods and information processing techniques, and the accompanying increase in computer computing power, has produced many mature algorithmic models in the field of deep learning, such as speech recognition, speech synthesis, computer graphics, and natural language processing. How to use these mature AI technologies to realize large-scale applications has gradually become a problem for researchers in these fields to explore and think about; based on this, “multimodality” is also increasingly talked about by researchers, and multimodal HCI mode will be a very important research direction in the field of AI in the future. The multimodal HCI model will be a very important research direction in the future AI field. Early speech recognition methods were simple template-matching methods, and the research direction was mainly for the recognition of isolated words and small vocabularies. After the 1980s, the research object of speech recognition shifted from isolated words to continuous words, and the research method gradually shifted to statistical template methods [3]. There were two very important modeling approaches at that time: acoustic models modeled using Gaussian mixture distribution–hidden Markov models (GMM–HMM) and language models based on N-element grammar modeling, and they dominated speech recognition for a long time. Although deep neural networks (DNN)–HMM-based speech recognition systems had achieved good recognition rates at that time, since neural networks actually work only on phoneme recognition, separate training is required for acoustic models, language models, and pronunciation models, and the aforementioned models have to be doubled if discriminative training is used, in addition to the fact that expertise is required to define pronunciation dictionaries and phoneme sets for specific languages, DNN The network structure and training process of traditional HMM-based speech recognition systems are very cumbersome [4]. Therefore, end-to-end structure-based modeling algorithms are a hot research topic for speech recognition tasks in recent years. The end-to-end structure-based speech recognition algorithm essentially solves the sequence-to-sequence mapping problem by mapping acoustic feature sequences to text sequences through a neural network structure, which does not require additional separate modeling of language models and pronunciation models and is much easier to train than traditional recognition frameworks. In this regard, this article proposes a speech recognition and intelligent translation method based on multimodal HCI system, aiming to improve the speech recognition effect and provide reliable reference for machine translation [5]. Speech recognition technology is an important means of achieving HCI. By converting speech signals into understandable text or commands, users can more conveniently interact with computers, and it can be applied in various fields. The research and development of speech recognition technology can improve the efficiency and comfort of people using computers and other intelligent devices. Intelligent translation refers to the process of using computer technology to convert text or spoken language from one language into another. The development of intelligent translation technology can effectively reduce communication barriers between different languages and promote cooperation and communication among humans in various fields such as economy, technology, and culture. A comparison between the proposed work and recent work on the subject of investigation is shown in Table 1.

Table 1

Comparison between proposed work and recent work on the subject of investigation

Contrast index Proposed work HMM Language model based on n-gram grammar modeling
Solve problems Solve the problem that only single-mode text images and text videos can be translated, and the translation accuracy is not high. Because neural network only works in phoneme recognition, it is necessary to train acoustic model, language model, and pronunciation model separately. Solve the problem of using specialized knowledge to determine the pronunciation dictionary and phoneme combination of a specific language.
Methods First, the network structure of speech recognition model in multi-channel HCI system is established, and the multi-head self-attention mechanism is constructed. Then, the voice wake-up function of AI is designed, and a multimodal machine translation model is established. Selective attention is added to obtain visual recognition of perceived text, and multi-modal gating fusion is carried out by decoder to realize the output of encoder translation results. Acoustic model based on Gaussian mixture distribution modeling Acoustic model based on Gaussian mixture distribution modeling
Innovation/Deficiency Modeling algorithm based on the end-to-end structure is a research hotspot in the field of speech recognition in recent years. The speech recognition algorithm based on the end-to-end structure maps the acoustic feature sequence to the text sequence through the neural network structure, which essentially solves the mapping problem from sequence to sequence, and it is easier to train than the traditional recognition framework without additional language model and pronunciation model. Translation robots can only translate single-mode text images and text videos. There is a problem of low translation accuracy.

2 Speech recognition technology under multimodal HCI system

2.1 Speech recognition network structure construction

The multimodal HCI system consists of three parts: input audio, intermediate link for cognitive and decision control, output audio and video, where the intermediate link for cognitive and decision control is the core algorithm part of the system, which combines technologies such as speech recognition, text dialogue, speech synthesis, and generating face expressions, and contains multimodal information such as speech, text, expressions, and faces. The input of the system is audio, and the output is video and audio, for which the speech recognition technology needs to be designed, and the specific implementation process is as follows [6].

The task of speech recognition converts the input audio into the corresponding text, which is the first link of the multimodal HCI system. In the task of speech recognition, a variety of speech data are usually used to evaluate the performance of the system. The number of voices used depends on the size of the data set. Generally speaking, in order to obtain reliable evaluation results, a certain amount of voice data is needed to test the system. Hundreds or thousands of speech samples can be used for testing to evaluate the accuracy of speech recognition and other indicators. Therefore, the accuracy of the model recognition will directly determine whether the output audio of the multimodal HCI system designed in this article can reasonably answer the input audio, which is an important key factor affecting the system performance [7]. The network structure of the model will be described below.

The network structure of the speech recognition model in this article is borrowed from the transformer model, which has superior performance in processing sequential data, and several modifications are made on this basis: first, the frame rate is reduced and the sequence length of acoustic features is reduced to improve the computational efficiency; second, the decoder does not completely use the teacher forcing technique during training, but sets a threshold between the predicted characters and the real characters, which determines whether the decoder uses predicted characters or real characters for the next moment of input; third, the minimum granularity of the units modeling the text in the decoder is subwords, and the network structure of speech recognition in this article is shown in Figure 1 [8].

Figure 1 
                  Network structure of speech recognition model under multimodal HCI system.
Figure 1

Network structure of speech recognition model under multimodal HCI system.

The model structure is a typical encoder-a-decoder-based network structure. In the multi-channel HCI system, the network structure of speech recognition model plays a vital role. This network structure needs strong feature extraction ability to effectively capture important information in speech signals; at the same time, it also needs to have the ability of context modeling to consider the influence of context information in speech signals on recognition results; in addition, the ability of multi-channel fusion is also important, so that the speech features from different channels can be integrated to improve the overall recognition performance and robustness; finally, in order to meet the requirements of real time, the network structure should have a low inference delay and can quickly process and recognize the input speech signal. By optimizing and improving the network structure, the accuracy and robustness of speech recognition can be improved to meet the needs of multi-channel HCI system. The encoder structure is a stack of N encoder layers; each layer contains two sub-layers: a multi-headed self-attentive mechanism and a simple fully connected feedforward network; the two sub-layers are connected by a residual network and then normalized by a layer [9]. The decoder structure is a stack of M decoder layers; each layer contains three sub-layers: a multi-headed self-attentive mechanism, an encoder–decoder multi-headed self-attentive mechanism, and a simple fully connected feedforward network; and each sub-layer is also connected using residuals and then layer normalization operations, and several important modules of the network structure are described below [10].

  1. Location code

    There is no convolution and loop operation in the network structure used in this article, and it is necessary to add some tagged position information to the input sequence, which can be the absolute or relative position information of the sequence. Therefore, the position encoding is added to the lowermost embedding layer of the codec structure, which has the same dimension as the input embedding d model , and the two can be added together [11]. The position encoding can be obtained by learning or calculated by a fixed calculation using sine and cosine functions with different frequencies, as shown below:

    (1) P E(pos , 2 i ) = sin p os d model ,

    (2) P E ( pos , 2 i + 1 ) = cos p os d model ,

    where p os denotes the position in the sequence, P E denotes the vector corresponding to the position p os , and d model denotes the dimension of the vector. The trigonometric function calculates the position encoding with a generative law that can be expected to have some extrapolation, and according to the trigonometric principle, the elements in P E ( p os + n ) can be represented by the elements in P E ( p os ) , so that the position encoding of each input feature is associated with [12].

  2. Layer normalization

Layer normalization is to normalize the output of the activation function to accelerate the convergence of the model so that all hidden units in the same layer of the network share the mean and variance, independent of the batch size.

2.2 Multi-headed self-attentive mechanism construction

Attention mechanism is a kind of data processing, and some scholars have previously achieved better results by applying attention mechanism to image classification tasks, and then, attention mechanism has gained wide attention in the field of deep learning [13,14]. Suppose that the attention mechanism is used to learn the context vector for an input sequence X (sequence length L and dimension d), a query vector Q is needed (the generation of this vector is determined by the specific task), and the attention score is obtained by calculating the correlation between the query vector and each input in the input sequence X through the scoring function (in this article, the attention score is calculated using dot product scaling), and then, the attention score is mapped into a softmax function. The softmax function maps the attention score into a probability distribution between (0, 1), and finally, the input information is weighted and summed to obtain the context vector. The attention calculation method mentioned earlier is called soft attention, and the attention mechanisms in this article use the soft attention method, as shown in the following expressions [15]:

(3) A ( Q , K , V ) = softmax Q KT d V ,

where A ( Q , K , V ) represents the attention mechanism that generates the query vector Q from the input sequence X, which is called the self-attentive mechanism. The self-attentive mechanism generally adopts the query one-key-one-value model, which maps the input sequence into Q, K, and V, respectively, by different weight vectors, where the attention score calculated by the scoring function is the correlation between Q and K. The attention mechanism that does multiple mappings of the query one-key-value vector is called the multi-headed attention mechanism, and the specific structure is shown in Figure 2 [16].

Figure 2 
                  Network structure diagram of the multi-headed attention mechanism.
Figure 2

Network structure diagram of the multi-headed attention mechanism.

Multi-attention mechanism is a common mechanism in neural networks, which is widely used in natural language processing and computer vision. Each attention header gets a set of attention weights by inner product of query and key. These attention weights reflect the correlation between the query and each key, i.e., the degree to which the query pays attention to the key. The final output of multi-head attention mechanism can be obtained by weighted summation of attention weights and corresponding values. Each attention head can learn different correlation patterns, thus capturing the diversity within the data. Specifically, in the network structure, input features are mapped into query, key, and value spaces through linear transformation, and these mapping matrices are learnable parameters of the network. The final output can be transmitted to the next layer of network for further processing or application.

The advantage of multi-head attention mechanism is that it can capture the information in the input data from different angles at the same time, and accurately weight and integrate the important information. By learning the weights between multiple attention heads, the network can flexibly adjust different attention modes and carefully interpret and deal with specific tasks.

The attention mechanism is characterized by focusing limited attention on important information, acquiring critical information and it is data-driven, with fewer training parameters than convolutional neural networks (CNNs) and recurrent neural networks while having the ability of CNN parallel processing. The self-attentive mechanism focuses too much on its own location while encoding the features at the current location, thus ignoring other information [17]. The multi-headed self-attentive mechanism is similar to the role of using multiple filters simultaneously in CNN, which can solve the problem that the self-attentive mechanism focuses too much on its own location features and can capture richer features, and the output of the attention layer collects the encoded features of all subspaces, making the model more expressive.

There are three forms of keys for the multi-headed self-attentive mechanism in the network structure of the model in this article. The query and key value in the self-attentive module of the encoder layer are derived from the output of the preceding layer. Similarly, in the decoder layer, the query and key value for the first self-attentive module are also sourced from the output of the preceding layer. However, due to the parallel nature of the attention mechanism, a mask is applied to the input to prevent the decoder from accessing irrelevant information during decoding. In the second self-attentive module of the decoder layer, the query is still obtained from the output of the previous layer, but the key values are now derived from the context vector features encoded by the encoder layer specifically for acoustic features [18].

Speech recognition is essentially a multi-classification problem, and the model uses a cross-entropy loss function, which is a loss function commonly used in multi-classification tasks. And the label smoothing strategy is used to reduce the confidence level of correctly classified samples to improve the adaptive ability of the model, with the following formulas [19]:

(4) H ( q , p ) = K = 1 K log P E ( p os ) ,

(5) q ' ( k ) = ( 1 ε ) K ,

where H represents the cross-entropy, ε represents the smoothing parameter, K represents the number of categories of token, ( q , p ) represents the model adaptive value, and q ( k ) represents the loss function. The aforementioned steps are organized to obtain the speech recognition process, as shown in Figure 3.

Figure 3 
                  Speech recognition process.
Figure 3

Speech recognition process.

As can be seen from Figure 3, based on the following four working principles, the speech recognition system is able to complete the recognition of speech. First, the speech recognition library and detection terminal are targeted and selected, and combined with anti-aliasing band-pass filtering technology, which can effectively eliminate the noise of individual speech differences, sampling equipment, and sampling environment; then, the speech acoustic parameters such as average energy and vibration peak and average over-zero rate are extracted, which can quickly and accurately reflect the phonetic characteristics of speech; then, the speech pattern database is established, and the link of language repetition training mainly is to let the speaker repeat the pronunciation, directly delete the redundant speech information one by one from the original speech samples, only keep part of the key speech data, and scientifically classify the key speech data according to the relevant scheme; finally, accurately output the relevant semantics of speech according to the speech similarity [20]. Through the aforementioned steps, the speech recognition network structure and the multi-headed attention mechanism network structure can be constructed and the loss function can be obtained to complete the design of speech recognition technology under the multimodal HCI system.

3 Intelligent translation technology under multimodal HCI system

3.1 AI voice wake-up function design

In order to implement the voice wake-up function, a process needs to be set up in the background to regularly monitor the environment around the device in real time and to detect the presence of the keyword required by the device in the signal, as shown in Table 2.

Table 2

AI voice wake-up process

Steps Process and instructions
01 Detection of process signals, hibernation for a specified length of time
02 Get PCM data and detect keywords
03 Determine whether to callback or not
04 Static detection of recording
05 Determine if the audio strength is less than the threshold value
06 Perform recording callback and return PCM audio data

As can be seen from Table 3, after completing the detection of the keyword signal, the detection callback or static detection is performed and the action is recorded. In the static detection of the recorded action, the process returns the recorded pulse code modulation (PCM) audio data to the main process and then continues to the next step [21]. Based on this, the data slicing process is performed on the captured pulse code modulated signal using the prime multistage dynamic labeling model. The machine translation steps of the HCI function are shown in Table 3.

Table 3

Machine translation process for HCI function

Steps Process and instructions
01 Read the machine translation model and select the pre-stored corresponding domain according to the user of the translation domain
02 Splitting the text into sentences
03 Search for matching translations on the search network

Since the number of states in the W n set increases exponentially with n, it would take a lot of time if the set is not pruned, so pruning is required. The pruning process is shown in the following:

For the identified source language sentence R 1 , R 2 , , R n , there exists a phrase model ( f , g , s , d ) , and the four elements in this model represent the phrase base, grammar model, distortion restrictions, and distortion parameters.

Assume that W n translates n words into m sets, and if an element of an m statement in W 2 consists of two words, it means that only two words are translated into the statement. For each state of W i , there is a translation state, and all possible states are added to the corresponding set, and the highest rated of them is finally returned [22].

Let β be the search parameter and p be the transfer parameter, and the resulting grammar model expression is shown in the following:

(6) g = n ext ( p , β ) .

After determining the search parameters, the pruning purpose is achieved after removing all the parameters in the set that do not satisfy equation (2):

(7) a ( g ) a ( p ) β .

When translating a source language sentence, the translation options of the source language sentence are first read, and then, the translation hypotheses are expanded from the small container to the large container. At each transition stage, if the difference between a score and the highest score in the container is greater than a threshold, the state decreases; if the state remains the same, all available translation options are expanded; if the old and new hypotheses are the same, the score is increased and the best translation result is to find the highest scoring translated statement in the largest container [23].

Voice wake-up function is one of the basic functions of multi-channel HCI system, which activates the system by recognizing specific voice commands and starts the subsequent interaction process. Perceptual visual representation of text is to transform text information into a representation form that computers can understand and process in a multimodal environment. The design and implementation of these two functions are helpful to improve the user experience and interaction effect of multi-channel HCI system.

3.2 Visual representation acquisition of perceptual text

There is a close relationship between designing AI voice wake-up function and obtaining visual representation of perceived text. As the entrance of the system, the voice wake-up function starts the system by recognizing specific voice commands; the visual representation of perceptual text provides a treatable representation form for text input in multimodal environment. The integration and collaborative design of these two functions is helpful to improve the efficiency and user experience of multi-channel HCI system and promote the development and application of HCI.

The main method to perceive the visual representation of the text is to collate the visual and textual information, align them, and find the similarity score between them to obtain the attention matrix, after which the visual representation weights are reassigned according to this matrix [24]. The specific steps are as follows:

  1. Transform the input text into word embeddings according to the word embedding table;

  2. Spatial visual features of images are obtained using the image pre-training model VCC19;

  3. Determining the visual feature dimension and the feature dimension of the text representation;

  4. If the dimensions of the two features are different, then a linear fully connected network is used for dimensional transformation so that the two dimensions are the same;

  5. Dot product processing and normalization of textual and visual features to obtain the attention matrix;

  6. Based on the attention matrix and spatial visual features from Steps (3) and (2), the visual representation of the perceived text is obtained;

  7. To encode and process textual and visual representations of perceptual text, two independent Transformer encoders are utilized, each responsible for one type of representation;

  8. Two text states are obtained, which are the hidden state and the visual hidden state;

  9. Reassign and fuse the hidden state weights.

3.3 Intelligent translation model construction and optimization

In the intelligent translation model in multimodal HCI systems, visual information can only play an auxiliary role, and if too much visual information is added to the model, it may make the redundant information in the model increase, thus making the model translation accuracy decrease. To solve this problem, this article proposes an intelligent translation model in multimodal HCI systems based on increasing the selection attention to replace the traditional feature summation by the important part of image features, which effectively reduces the interference of redundant information.

Assuming that E text [ e 1 , e 2 , , e n ] and F img [ f 1 , f 2 , , f n ] represent the textual representation and the spatial visual representation, respectively, the attention matrix can be obtained by finding the similarity between them, and the specific matrix expression is shown in the following:

(8) M atrix att = E text × F img T ,

where M atrix att represents the attention matrix, where each element represents the similarity of a word to a visual feature. The dot product of the textual representation e i and the visual representation f i is calculated to obtain M atrix att . Each element is selected by Gumbel sigmoid to obtain the region of the visual feature that is more relevant to the current one. Once the attention matrix is obtained, it can be exponentially normalized to a probability distribution where the similarity score does not sum to 1 and then multiplied with each element of the sampling result matrix M atrix sel to obtain the multimodal attention matrix M atrix mm , which is calculated as shown in the following:

(9) M atrix mm = M atrix sel × soft max ( M atrix att ) .

To further enhance the modeling capability of selective attention, a multi-headed attention mechanism can also be added to this attention, as shown in the following expression:

(10) M HGA ( Q , K , V ) = C oncat ( G A 1 , G A 2 , , G A h ) W 0 ,

where M HGA and h denote the multi-head selective attention and the number of this attention head, respectively; G A 1 , G A 2 , , G A h both denote the sub-attention module, which is processed by splicing and linear transformation to obtain the final output [25]. Based on the improvement of the aforementioned attention matrix, the encoder of the multimodal translation model based on selective attention is obtained, as shown in Figure 4.

Figure 4 
                  Encoder structure of multimodal translation model based on selective attention.
Figure 4

Encoder structure of multimodal translation model based on selective attention.

Through the aforementioned steps, the optimization and design of multimodal machine translation model can be completed, and the encoder can be obtained to realize the machine translation function. By combining the machine translation-related contents in this section with the speech recognition technology mentioned earlier, the design of speech recognition and intelligent translation technology under multimodal HCI system is completed.

4 Testing and analysis

4.1 Test preparation

In order to prove that the speech recognition and intelligent translation techniques under the multimodal HCI system proposed in this article are better than the conventional speech recognition and intelligent translation techniques in terms of translation effect, after the theoretical part of the design is completed, an experimental session is constructed to test the evaluation effect of the method in this article. In order to improve the reliability of the experimental results, in addition to the method of this article, two conventional methods are selected as the experimental control group, which are the speech recognition and intelligent translation technology based on fuzzy algorithm and the speech recognition and intelligent translation technology based on support vector machine (SVM).

To obtain better experimental results, the experimental data were chosen to be obtained from the Multi30K dataset. This dataset contains 30,000 text-image data pairs with a total of 30 000 entries. Among them, the training set, test set, and validation set are 25,000, 4,200, and 800 items, respectively. The text-a-video data was derived from the VATEX dataset with a total of 42,000 entries. The data percentages of training set, test set, and validation set are 60, 20, and 20%, respectively, and the classification results of the test set are shown in Table 4.

Table 4

Test set classification

Types of sentences Number of words Number of sentences
Simple sentences N < 8 458
Normal sentences 8 ≤ N < 18 368
Complex sentences N > 18 178

To ensure the experimental effect, different parameters will be set according to the difference of datasets for the text-image and text-video machine translation tasks. The numbers of layers of encoder and decoder are set to 4 and 5 for text-images and text-video machine translation tasks, respectively; the word vector dimension is set to 128 and 256, respectively; the self-attention and multimodal attention between encoder and decoder are set to 4, and the attention is set to 4 and 8, respectively; the initial learning rate is set to 0.001 Optimizer is selected as Adam, Batch Size.

4.2 Analysis of test results

The comparison criterion chosen for this experiment is the translation performance of the method, and the specific measure is the BLUE index, which is mainly used to evaluate the accuracy of translation, and the basic principle is to find out the degree of N-gram contribution between the generated translation and the reference translation, and the specific calculation formula is shown in the following formula:

(11) B LUE = i k min h k ( c i ) B P ,

where B P represents the penalty factor, i and k represent the kth N-gram fragment and the ith sentence, respectively, h k represents the current utterance, and c i represents the standard reference translation.

Analyzing the data in Table 5, we can see that compared with the other two conventional models, the BLUE value of the machine translation model under multimodal HCI system proposed in this article is higher, which proves that the model proposed in this article has higher translation accuracy and better model performance.

Table 5

Comparison results of BLUE values for different translation models

Number of tests Fuzzy algorithms SVMs Multimodal HCI
01 37.5 39.2 42.6
02 36.8 40.1 43.7
03 38.2 38.5 44.6
04 36.4 38.4 43.2
05 35.3 39.7 46.1

Choose accuracy as the evaluation index, the higher the accuracy value, the better the representation effect, and its calculation formula is as follows:

(12) F PN = F P F P + F N × 100 %,

where F PN represents the accuracy of speech recognition and intelligent translation, F P represents the number of correctly predicted samples, and F N represents the total number of samples. Using formula (12), the sentences in the experimental dataset are translated by fuzzy algorithm, SVM, and multi-channel HCI, respectively, and the accuracy of translation results is shown in Figure 5.

Figure 5 
                  Comparison results of translation accuracy of three methods.
Figure 5

Comparison results of translation accuracy of three methods.

As can be seen from Figure 5, compared with fuzzy algorithm and SVM method, the accuracy of multi-channel HCI method is higher, with the highest value of 99%, which indicates that the research method has better speech recognition and intelligent translation effect.

Use confusion matrix to verify the effect of speech recognition and intelligent translation in multimodal HCI system. Confusion matrix, also known as error matrix, is a standard format for expressing accuracy evaluation, usually expressed in the form of matrix with m rows and n columns, specifically as follows:

(13) P E = p 11 p 12 p 1 m p 21 p 22 p 2 m p n 1 p n 2 p n m .

Specifically, each column of the confusion matrix represents the prediction category, and the total number of each column represents the number of data predicted as this category; each row represents the real category of data, and the total number of data in each row represents the number of data instances in this category. Through the confusion matrix, we can intuitively understand the performance of speech recognition and intelligent translation. Under the application of different methods, three test samples are selected, and the samples are identified by mistake and missed speech, and their numbers are counted, and then, the errors are calculated. Finally, the results are compared, as shown in Table 6.

Table 6

Error comparison results of different methods

Different methods Test sample Error (%) Average value (%)
Multi-modal HCI method Test sample 1 1.2 1.5
Test sample 2 2.1
Test sample 3 1.2
Fuzzy algorithm Test sample 1 6.9 7.3
Test sample 2 7.3
Test sample 3 7.7
SVM method Test sample 1 7.5 7.7
Test sample 2 6.7
Test sample 3 8.9

As can be seen from Table 6, the error of multimodal HCI method on test sample 1 is 1.2%, and the average error is 1.5%. The error of fuzzy algorithm on test sample 1 is 6.9%, and the average error is 7.3%. The error of SVM method on test sample 1 is 7.5%, and the average error is 7.7%. In the test sample 1, the error of multimodal HCI method is the lowest, while the error of SVM method is the highest. In terms of average value, multi-modal HCI method is still the most accurate, and the average error of fuzzy algorithm and SVM method is high; through the aforementioned analysis, we can know that the multimodal HCI method performs best in these test samples, with higher accuracy and stability; The performance of fuzzy algorithm on test samples is inferior to that of multimodal HCI method, but better than that of SVM method. SVM performs worst among these test samples and has a high error rate. This shows that the research method is more accurate and comprehensive, and the recognition effect is better.

To sum up, the speech recognition and intelligent translation methods in this multi-modal HCI system have achieved better performance. By evaluating the BLUE value, the similarity between the translation results and the manual reference translation can be quantified. It has higher BLUE value and accuracy, which shows that it has better translation accuracy and model performance, and provides a more reliable language processing tool for cross-cultural communication and cooperation.

5 Conclusion

In the multi-modal HCI system, speech recognition and intelligent translation are two key technologies, and their combination can provide strong support for cross-cultural communication and cooperation. This article discusses the research of speech recognition and intelligent translation in multimodal HCI system and draws the following conclusions: the model proposed in this article has higher translation accuracy and better model performance. It shows that the model in this article has a higher BLUE value than the traditional translation model, which verifies its advantages in translation accuracy and model performance.

However, there are still some challenges and room for improvement. First of all, speech recognition technology is still affected by background noise and speaker changes, and further research is needed to improve robustness and reliability. Second, in the field of intelligent translation, although the model has achieved good results in a multimodal environment, it still needs to solve some specific problems between languages, such as structural differences and cultural differences.

  1. Funding information: Shaoxing Higher Education Teaching Reform Research Program (SXSJG202402).

  2. Author contributions: All authors participated in some part of the work for this article. In the investigation, DH proposed the idea and conceived the design, carried out the simulation, and wrote the original draft; SX analyzed and discussed the results and reviewed and edited the manuscript. All authors have read and agreed to the published version of the manuscript.

  3. Conflict of interest: The authors declare no conflict of interest.

  4. Data availability statement: Data sharing is not applicable to this article as no datasets were generated or analysed during the current study.

References

[1] Badrinath S, Balakrishnan H. Automatic speech recognition for air traffic control communications. Transp Res Rec. 2022;2676(1):798–810.10.1177/03611981211036359Search in Google Scholar

[2] Zeng T, Yang X, Wan Y, Mao Y, Liu Z. Effectiveness assessment of improvement measures in physical protection system monitoring center. Kerntechnik. 2021;86(1):33–8.10.1515/KERN-2020-0026Search in Google Scholar

[3] Song T, Zhao H, Liu Z, Liu H, Hu Y, Sun D. Intelligent human hand gesture recognition by local-global fusing quality-aware features. Future Gener Comput Syst. 2021;115(7043):298–303.10.1016/j.future.2020.09.013Search in Google Scholar

[4] Roda-Sanchez L, Olivares T, Garrido-Hidalgo C, de la Vara JL, Fernández-Caballero A. Human-robot interaction in industry 4.0 based on an internet of things real-time gesture control system. Integr Comput Eng. 2021;28(2):1–17.10.3233/ICA-200637Search in Google Scholar

[5] Porcheron M, Fischer JE, Reeves S. Pulling back the curtain on the wizards of oz. Proc ACM Human-Comput Interact. 2021;4(CSCW3):1–22.10.1145/3432942Search in Google Scholar

[6] Zhang Y. Interactive intelligent teaching and automatic composition scoring system based on linear regression machine learning algorithm. J Intell Fuzzy Syst. 2021;40(2):2069–81.10.3233/JIFS-189208Search in Google Scholar

[7] Zhang D. Intelligent recognition of dance training movements based on machine learning and embedded system. J Intell Fuzzy Syst. 2021;1:1–13.10.3233/JIFS-189954Search in Google Scholar

[8] Jasim M, Khaloo P, Wadhwa S, Zhang AX, Sarvghad A, Mahyar N. Community click: capturing and reporting community feedback from town halls to improve inclusivity. Proc ACM Human-Comput Interact. 2021;4(CSCW3):1–32.10.1145/3432912Search in Google Scholar

[9] Yang B, Xia X, Wang S, Ye L. Development of flight simulation system based on leap motion controller. Proc Comput Sci. 2021;183(2):794–800.10.1016/j.procs.2021.02.131Search in Google Scholar

[10] Jing W, Tao H, Rahman MA, Kabir MN, Yafeng L, Zhang R, et al. RERS-CC: Robotic facial recognition system for improving the accuracy of human face identification using HRI. Work. 2021;68(7):1–12.10.3233/WOR-203426Search in Google Scholar PubMed

[11] Yu J, Ji H, Song Q, Zhou L. Design and implementation of business access control in new generation power grid dispatching and control system. Proc Comput Sci. 2021;183(22):761–7.10.1016/j.procs.2021.02.126Search in Google Scholar

[12] Carlos Alberto PJ, Sonia Karina PJ, Francisca Irene SA, Adrielly Nahomee RÁ. Waste reduction in printing process by implementing a video inspection system as a human machine interface. Proc Comput Sci. 2021;180:79–85.10.1016/j.procs.2021.01.131Search in Google Scholar

[13] Sha Y, Feng T, Xiong X, Yang T. Designing online psychological consultation expert system using human-computer interaction. Mob Inf Syst. 2021;2021(1):1–12.10.1155/2021/6458924Search in Google Scholar

[14] Su KW, Chiu PC, Lin TH. Establishing a blockchain online travel agency with a human-computer interaction perspective. J Hosp Tour Technol. 2022;13(3):559–72.10.1108/JHTT-01-2021-0038Search in Google Scholar

[15] Mitchell EG, Maimone R, Cassells A, Tobin JN, Davidson P, Smaldone AM, et al. Automated vs. human health coaching: exploring participant and practitioner experiences. Proc ACM Human-Comput Interact. 2021;5(CSCW1):1–37.10.1145/3449173Search in Google Scholar PubMed PubMed Central

[16] Zhang H. Voice keyword retrieval method using attention mechanism and multimodal information fusion. Sci Program. 2021;2021(8):1–11.10.1155/2021/6662841Search in Google Scholar

[17] Yuan Q, Wang R, Pan Z, Xu S, Gao J, Luo T. A survey on human-computer interaction in spatial augmented reality. J Comput Des Comput Graph. 2021;33(3):321–32.10.3724/SP.J.1089.2021.18445Search in Google Scholar

[18] Sreekanth NS, Narayanan NK. Multimodal human computer interaction with context dependent input modality suggestion and dynamic input ambiguity resolution. Int J Eng Trends Technol. 2021;69(5):152–65.10.14445/22315381/IJETT-V69I5P222Search in Google Scholar

[19] Evers K, Chen S. Effects of automatic speech recognition software on pronunciation for adults with different learning styles. J Educ Comput Res. 2021;59(4):669–85.10.1177/0735633120972011Search in Google Scholar

[20] Alhumsi MH, Belhassen S. The challenges of developing a living Arabic phonetic dictionary for speech recognition system: A literature review. Adv J Soc Sci. 2021;8(1):164–70.10.21467/ajss.8.1.164-170Search in Google Scholar

[21] Kempfle JS, Panda A, Hottin M, Vinik K, Kozin ED, Ito CJ, et al. Effect of powered air-purifying respirators on speech recognition among health care workers. Otolaryngol-Head Neck Surg. 2021;164(1):87–90.10.1177/0194599820945685Search in Google Scholar PubMed

[22] Ji YJ, Bahng J, Lee JH. Efficacy of a closed-set auditory training protocol on speech recognition of adult hearing aid users. Korean J Otorhinolaryngol – Head Neck Surg. 2021;64(2):70–6.10.3342/kjorl-hns.2019.00696Search in Google Scholar

[23] Folkeard P, Eeckhoutte MV, Levy S, Dundas D, Abbasalipour P, Glista D, et al. Detection, speech recognition, loudness, and preference outcomes with a direct drive hearing aid: Effects of band width. Trends Hearing. 2021;25(4):8–13.10.1177/2331216521999139Search in Google Scholar PubMed PubMed Central

[24] Sun Z, Tang P. Automatic communication error detection using speech recognition and linguistic analysis for proactive control of loss of separation. Transp Res Rec. 2021;2675(5):1–12.10.1177/0361198120983004Search in Google Scholar

[25] Kumar LA, Renuka DK, Rose SL, Shunmuga priya MC, Wartana IM. Deep learning based assistive technology on audio visual speech recognition for hearing impaired. Int J Cognit Comput Eng. 2022;3:24–30.10.1016/j.ijcce.2022.01.003Search in Google Scholar

Received: 2023-10-11
Accepted: 2024-04-15
Published Online: 2024-09-03

© 2024 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Research Articles
  2. A study on intelligent translation of English sentences by a semantic feature extractor
  3. Detecting surface defects of heritage buildings based on deep learning
  4. Combining bag of visual words-based features with CNN in image classification
  5. Online addiction analysis and identification of students by applying gd-LSTM algorithm to educational behaviour data
  6. Improving multilayer perceptron neural network using two enhanced moth-flame optimizers to forecast iron ore prices
  7. Sentiment analysis model for cryptocurrency tweets using different deep learning techniques
  8. Periodic analysis of scenic spot passenger flow based on combination neural network prediction model
  9. Analysis of short-term wind speed variation, trends and prediction: A case study of Tamil Nadu, India
  10. Cloud computing-based framework for heart disease classification using quantum machine learning approach
  11. Research on teaching quality evaluation of higher vocational architecture majors based on enterprise platform with spherical fuzzy MAGDM
  12. Detection of sickle cell disease using deep neural networks and explainable artificial intelligence
  13. Interval-valued T-spherical fuzzy extended power aggregation operators and their application in multi-criteria decision-making
  14. Characterization of neighborhood operators based on neighborhood relationships
  15. Real-time pose estimation and motion tracking for motion performance using deep learning models
  16. QoS prediction using EMD-BiLSTM for II-IoT-secure communication systems
  17. A novel framework for single-valued neutrosophic MADM and applications to English-blended teaching quality evaluation
  18. An intelligent error correction model for English grammar with hybrid attention mechanism and RNN algorithm
  19. Prediction mechanism of depression tendency among college students under computer intelligent systems
  20. Research on grammatical error correction algorithm in English translation via deep learning
  21. Microblog sentiment analysis method using BTCBMA model in Spark big data environment
  22. Application and research of English composition tangent model based on unsupervised semantic space
  23. 1D-CNN: Classification of normal delivery and cesarean section types using cardiotocography time-series signals
  24. Real-time segmentation of short videos under VR technology in dynamic scenes
  25. Application of emotion recognition technology in psychological counseling for college students
  26. Classical music recommendation algorithm on art market audience expansion under deep learning
  27. A robust segmentation method combined with classification algorithms for field-based diagnosis of maize plant phytosanitary state
  28. Integration effect of artificial intelligence and traditional animation creation technology
  29. Artificial intelligence-driven education evaluation and scoring: Comparative exploration of machine learning algorithms
  30. Intelligent multiple-attributes decision support for classroom teaching quality evaluation in dance aesthetic education based on the GRA and information entropy
  31. A study on the application of multidimensional feature fusion attention mechanism based on sight detection and emotion recognition in online teaching
  32. Blockchain-enabled intelligent toll management system
  33. A multi-weapon detection using ensembled learning
  34. Deep and hand-crafted features based on Weierstrass elliptic function for MRI brain tumor classification
  35. Design of geometric flower pattern for clothing based on deep learning and interactive genetic algorithm
  36. Mathematical media art protection and paper-cut animation design under blockchain technology
  37. Deep reinforcement learning enhances artistic creativity: The case study of program art students integrating computer deep learning
  38. Transition from machine intelligence to knowledge intelligence: A multi-agent simulation approach to technology transfer
  39. Research on the TF–IDF algorithm combined with semantics for automatic extraction of keywords from network news texts
  40. Enhanced Jaya optimization for improving multilayer perceptron neural network in urban air quality prediction
  41. Design of visual symbol-aided system based on wireless network sensor and embedded system
  42. Construction of a mental health risk model for college students with long and short-term memory networks and early warning indicators
  43. Personalized resource recommendation method of student online learning platform based on LSTM and collaborative filtering
  44. Employment management system for universities based on improved decision tree
  45. English grammar intelligent error correction technology based on the n-gram language model
  46. Speech recognition and intelligent translation under multimodal human–computer interaction system
  47. Enhancing data security using Laplacian of Gaussian and Chacha20 encryption algorithm
  48. Construction of GCNN-based intelligent recommendation model for answering teachers in online learning system
  49. Neural network big data fusion in remote sensing image processing technology
  50. Research on the construction and reform path of online and offline mixed English teaching model in the internet era
  51. Real-time semantic segmentation based on BiSeNetV2 for wild road
  52. Online English writing teaching method that enhances teacher–student interaction
  53. Construction of a painting image classification model based on AI stroke feature extraction
  54. Big data analysis technology in regional economic market planning and enterprise market value prediction
  55. Location strategy for logistics distribution centers utilizing improved whale optimization algorithm
  56. Research on agricultural environmental monitoring Internet of Things based on edge computing and deep learning
  57. The application of curriculum recommendation algorithm in the driving mechanism of industry–teaching integration in colleges and universities under the background of education reform
  58. Application of online teaching-based classroom behavior capture and analysis system in student management
  59. Evaluation of online teaching quality in colleges and universities based on digital monitoring technology
  60. Face detection method based on improved YOLO-v4 network and attention mechanism
  61. Study on the current situation and influencing factors of corn import trade in China – based on the trade gravity model
  62. Research on business English grammar detection system based on LSTM model
  63. Multi-source auxiliary information tourist attraction and route recommendation algorithm based on graph attention network
  64. Multi-attribute perceptual fuzzy information decision-making technology in investment risk assessment of green finance Projects
  65. Research on image compression technology based on improved SPIHT compression algorithm for power grid data
  66. Optimal design of linear and nonlinear PID controllers for speed control of an electric vehicle
  67. Traditional landscape painting and art image restoration methods based on structural information guidance
  68. Traceability and analysis method for measurement laboratory testing data based on intelligent Internet of Things and deep belief network
  69. A speech-based convolutional neural network for human body posture classification
  70. The role of the O2O blended teaching model in improving the teaching effectiveness of physical education classes
  71. Genetic algorithm-assisted fuzzy clustering framework to solve resource-constrained project problems
  72. Behavior recognition algorithm based on a dual-stream residual convolutional neural network
  73. Ensemble learning and deep learning-based defect detection in power generation plants
  74. Optimal design of neural network-based fuzzy predictive control model for recommending educational resources in the context of information technology
  75. An artificial intelligence-enabled consumables tracking system for medical laboratories
  76. Utilization of deep learning in ideological and political education
  77. Detection of abnormal tourist behavior in scenic spots based on optimized Gaussian model for background modeling
  78. RGB-to-hyperspectral conversion for accessible melanoma detection: A CNN-based approach
  79. Optimization of the road bump and pothole detection technology using convolutional neural network
  80. Comparative analysis of impact of classification algorithms on security and performance bug reports
  81. Cross-dataset micro-expression identification based on facial ROIs contribution quantification
  82. Demystifying multiple sclerosis diagnosis using interpretable and understandable artificial intelligence
  83. Unifying optimization forces: Harnessing the fine-structure constant in an electromagnetic-gravity optimization framework
  84. E-commerce big data processing based on an improved RBF model
  85. Analysis of youth sports physical health data based on cloud computing and gait awareness
  86. CCLCap-AE-AVSS: Cycle consistency loss based capsule autoencoders for audio–visual speech synthesis
  87. An efficient node selection algorithm in the context of IoT-based vehicular ad hoc network for emergency service
  88. Computer aided diagnoses for detecting the severity of Keratoconus
  89. Improved rapidly exploring random tree using salp swarm algorithm
  90. Network security framework for Internet of medical things applications: A survey
  91. Predicting DoS and DDoS attacks in network security scenarios using a hybrid deep learning model
  92. Enhancing 5G communication in business networks with an innovative secured narrowband IoT framework
  93. Quokka swarm optimization: A new nature-inspired metaheuristic optimization algorithm
  94. Digital forensics architecture for real-time automated evidence collection and centralization: Leveraging security lake and modern data architecture
  95. Image modeling algorithm for environment design based on augmented and virtual reality technologies
  96. Enhancing IoT device security: CNN-SVM hybrid approach for real-time detection of DoS and DDoS attacks
  97. High-resolution image processing and entity recognition algorithm based on artificial intelligence
  98. Review Articles
  99. Transformative insights: Image-based breast cancer detection and severity assessment through advanced AI techniques
  100. Network and cybersecurity applications of defense in adversarial attacks: A state-of-the-art using machine learning and deep learning methods
  101. Applications of integrating artificial intelligence and big data: A comprehensive analysis
  102. A systematic review of symbiotic organisms search algorithm for data clustering and predictive analysis
  103. Modelling Bitcoin networks in terms of anonymity and privacy in the metaverse application within Industry 5.0: Comprehensive taxonomy, unsolved issues and suggested solution
  104. Systematic literature review on intrusion detection systems: Research trends, algorithms, methods, datasets, and limitations
Downloaded on 5.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jisys-2023-0192/html
Scroll to top button