[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Research on Motion Control and Wafer-Centering Algorithm of Wafer-Handling Robot in Semiconductor Manufacturing
Previous Article in Journal
Review of Random Fiber Lasers for Optical Fiber Sensors
Previous Article in Special Issue
LLM Adaptive PID Control for B5G Truck Platooning Systems
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Interactive Framework of Cross-Lingual NLU for In-Vehicle Dialogue

School of Artificial Intelligence and Big Data, Hefei University, Hefei 230061, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(20), 8501; https://doi.org/10.3390/s23208501
Submission received: 17 August 2023 / Revised: 26 September 2023 / Accepted: 10 October 2023 / Published: 16 October 2023

Abstract

:
As globalization accelerates, the linguistic diversity and semantic complexity of in-vehicle communication is increasing. In order to meet the needs of different language speakers, this paper proposes an interactive attention-based contrastive learning framework (IABCL) for the field of in-vehicle dialogue, aiming to effectively enhance cross-lingual natural language understanding (NLU). The proposed framework aims to address the challenges of cross-lingual interaction in in-vehicle dialogue systems and provide an effective solution. IABCL is based on a contrastive learning and attention mechanism. First, contrastive learning is applied in the encoder stage. Positive and negative samples are used to allow the model to learn different linguistic expressions of similar meanings. Its main role is to improve the cross-lingual learning ability of the model. Second, the attention mechanism is applied in the decoder stage. By articulating slots and intents with each other, it allows the model to learn the relationship between the two, thus improving the ability of natural language understanding in languages of the same language family. In addition, this paper constructed a multilingual in-vehicle dialogue (MIvD) dataset for experimental evaluation to demonstrate the effectiveness and accuracy of the IABCL framework in cross-lingual dialogue. With the framework studied in this paper, IABCL improves by 2.42% in intent, 1.43% in slot, and 2.67% in overall when compared with the latest model.

1. Introduction

The intelligent development of in-vehicle dialogue systems is one of the main goals to improve the functional convenience and driving safety of automobile cockpits [1,2,3,4,5]. Natural language understanding (NLU) is crucial for the implementation of voice control functions in in-vehicle dialogue systems, which mainly consists of two tasks: intent detection and slot filling. Intent detection means that when the user inputs a command to the dialogue systems, the computer determines whether the question is to turn on the air conditioner, listen to music, query the navigation location, etc. Slot filling is to determine the specific content of the intent based on the intent detection. For example, if the user inputs a navigation command to ‘go to the airport’, the dialogue system first determines that the intent of this sentence is to navigate, and the specific navigation location is the airport.
Since 2019, with the development of pretrained models in the field of natural language processing (NLP) and the improvement of representation capabilities, researchers have proposed a variety of joint-training frameworks [6,7,8,9,10,11,12,13] based on pretrained models to improve the ability of cross-lingual NLU. These frameworks have been applied in the industrial domain [6], cross-lingual interactive chatbots [7], cross-lingual e-banking domain [8], restaurant customer service domain [9], airline ticket purchasing customer service domain [10,11,12,13], and so on. These studies [6,7,8,9,10,11,12,13] show that the representation capability of pretrained models and the application of cross-lingual transfer learning techniques can play an important role in multilingual NLU in different domains.
To summarize, researchers have carried out applied research on cross-lingual NLU in many fields, but so far, there is no research on cross-lingual NLU tasks of a dialogue system in the field of in-vehicle dialogue. According to market research, the existing cross-lingual NLU in the vehicle field mainly adopts the rule-based method [14]. However, with the addition of some new knowledge and new terms, the rule-based method needs to spend a lot of manpower and material resources to adjust. In addition, there are currently private datasets in the field of in-vehicle dialogue, so it is necessary to build a dataset for an in-vehicle dialogue system, which can be shared and used by other researchers to promote the development of an in-vehicle dialogue system.
This paper focuses on the in-vehicle domain, attaining multilingual NLU and providing a set of datasets for multilingual in-vehicle dialogue. The contributions of this paper are mainly as follows:
(1)
This paper provided a set of datasets of NLU tasks of an in-vehicle dialogue system under cross-lingual scenarios, including Chinese, Arabic, Japanese, and English. The intent and slots of the datasets are annotated according to the annotation specification of the currently published MultiATIS++ [10] dataset.
(2)
It proposed a cross-lingual interactive framework for in-vehicle human–computer dialog systems. The framework employs an end-to-end architecture for joint modeling, using pretrained models and incorporating contrastive learning to communicate the relationship between intent recognition and slot filling tasks through interactive attention mechanisms in downstream tasks.
(3)
It validated an IABCL framework on self-built MIvD datasets and public datasets. To the best of our knowledge, this is the first time to evaluate the model work on the joint task of intent detection and slot filling in in-vehicle dialogue systems. The experimental results show that the attention-based interactive framework achieves remarkable results in cross-lingual intent detection and slot filling.

2. Related Work on Cross-Lingual NLU

2.1. In-Vehicle Dialogue NLU

Under the impetus of economic globalization, the number of automobile users has been growing gradually and the languages they use are different, so it is an urgent problem for NLU in the automobile field to correctly identify the intent detection and slot filling of the user’s expressions in cross-lingual environments.
This paper investigates cross-lingual NLU tasks in other domains. For example, in the industrial domain, in order to perform multilingual intent classification under low resources, Khalil et al. [6] constructed an industrial intent dataset and explored different methods for multilingual intent detection in industrial environments. The presented methods [6] verified that the source language can be well migrated to the target language based on the machine translation model and the multilingual pretraining model. In 2020, in the cross-lingual interactive chatbot domain, in order to better utilize cross-lingual transfer learning techniques to improve the accuracy of chatbots in multilingual prediction, especially for the COVID-19-related problem, Arora et al. [7] proposed a joint model that is structured for simultaneous intent detection and slot filling, and compared the effectiveness of different pretrained models in the cross-lingual transfer learning task. In addition, the authors provide a multilingual dataset, M-CID, containing English, Spanish, French, and German. In 2021, Daniela et al. [8] combined machine translation with a pretrained model for an intent detection task in the e-banking domain and obtained excellent results, and this article also contributed a new data resource called MIN DS-14. In the restaurant customer service domain, in order to reduce the cost of cross-lingual NLU, Liu et al. [9] proposed an attention-informed mixed-language training method. This mixed language training method uses a very small number of task-related parallel word pairs to generate sentences in a hybrid language. It outperforms the then zero-sample adaptation methods in dialogue state tracking and natural language understanding, and is less costly. In the field of ticketing customer service in the aircraft aviation domain, Xu et al. [10] proposed a new end-to-end model in order to address slot alignment and recognition in cross-lingual migration, which outperforms the traditional simple tag mapping method using fast-align on most languages, and the authors also released a new multilingual NLU dataset, MultiATIS++. Jack et al. [11], in the field of aircraft aviation, described a new multilingual NLU task that simultaneously translates input text, outputs text with slots, categorizes intents, recognizes the input language, and improves on the MultiATIS++ dataset using the multilingual BART model. In the same area, in 2020, Qin et al. [12] proposed a new approach to improve the language neutrality of Transformer-based language models (e.g., mBERT) by switching data through multilingual codes for fine-tuning downstream NLU tasks. Compared with machine translation, multilingual code switching allows for multilingual text characterization at a finer granularity, such as the sentence, phrase, or word level, while the authors released a new manually annotated tweet dataset for intent detection and slot filling in English and Haitian. In 2022, to address the problem of sequential alignment between multilingual transitions in the field of aircraft aviation, Qin et al. [13] proposed a contrastive learning framework. The framework achieves excellent performance on ticketing customer service in the aviation domain and successfully brings the representations of cross-lingual similar sentences closer together.

2.2. Cross-Lingual Transfer Learning

In the field of single-language dialogue system, NLU based on deep learning has achieved remarkable results [15]. However, in the field of multilingualism, especially in low-resource languages, the lack of a large amount of training data makes the model of intent detection and slot filling not better aligned between different languages, so its recognition effect in the cross-lingual domain is not significant. In recent years, some researchers have adopted the concept of transfer learning to train NLU models in resource-rich languages, such as English and Chinese, and adopted cross-lingual transfer learning strategies to apply them to other low-resource languages [16].
Cross-lingual transfer learning is a machine learning technique that allows knowledge learned in one language to be applied to another. The basic idea is to transfer the knowledge of the language model of one or more source languages into the language model of the target language [17]. There are two main approaches to cross-lingual transfer learning: one based on machine translation and the other based on a multilanguage model. The approach based on machine translation is to translate resource-rich data into a resource-poor target language and then train the model of the target language. For example, in 2010, Lefevre et al. [18] proposed a method to complete the intent detection task in the target language through machine translation given the source language and its semantic annotations. In 2017, Adams et al. [19] proposed a method using cross-lingual word vectors to solve the natural language processing problem of low-resource languages. It uses bilingual dictionaries to link monolingual corpora of two different languages to learn word embeddings in both languages in a common vector space, thereby helping to solve problems such as data scarcity and lack of labeled data in low-resource languages, improving the performance of natural language processing tasks. Andrew et al. [20] proposed a method of MOM (mixed orthographic mapping) to overcome the spelling differences between English and Japanese and realize the application of cross-lingual transfer learning in Japanese-named entity recognition. Another approach based on multilingual models is usually taking a pretrained model with some modifications and fine-tuning. In 2020, Qin et al. [12] proposed a multilingual code-switching data enhancement framework to fine-tune mBERT to align the representation of the source language and multiple target languages, improving the performance of cross-lingual NLU tasks. Qin et al. [13] proposed to use contrast learning to narrow the representation of cross-lingual similar sentences in mBERT to solve the word order alignment problem between multilanguage code transitions and achieved outstanding performance.

2.3. Joint Learning Frameworks for Intent Detection and Slot Filling

Joint learning of intent detection and slot filling is the mainstream model training method in the field of NLU at present. Parallel architecture is the earliest joint training method, which treats the two tasks as parallel subtasks instead of serial ones when processing intent detection and slot filling. Li et al. [21] constructed a parallel architecture of intent detection and slot filling. The architecture adopted bidirectional LSTM as the embedding layer, introduced attention-based classifiers for intent detection, and modeled slot filling with a conditional random field (CRF). Schuster et al. [22] used a self-attentional classifier as intent detection in the decoding layer, filled the CRF modeling slot, and compared some multilanguage models in the encoder. The advantage of a parallel architecture is that both intent detection and slot filling can be performed simultaneously, thus improving the efficiency and performance of the model. However, the disadvantage is that it ignores the strong correlation between the two tasks, and cannot supervise each other in the process of transfer learning, so the overall accuracy of the model is affected.
Aiming at the bidirectional interaction between intent and slot, Zhang et al. [23] proposed a dual-model architecture in 2016 to consider the cross-impact between intent and slot by using two related bidirectional LSTMS. Wang et al. [24] proposed to share parameters of RNN (Joint ID and SF) to learn the correlation between intent and slot. Compared with parallel architecture, the results of both intent detection and slot filling can be significantly improved with the same number of parameters. Liu et al. [25] then introduced a shared encoder–decoder framework with an attention mechanism, and introduced attention mechanisms into RNN models that provide additional information for intent classification and slot filling. Compared with shared RNN parameters, the bidirectional interaction for intent detection and slot filling has improved the results. In 2019, Chen et al. [26] proposed a BERT-based joint intent classification and slot filling model, which uses a universal large-scale language to represent the model BERT compared with the RNN model, reducing the training time. In 2021, Zhou et al. [27] introduced a new parallel interaction network (PIN) in which the interaction between intent detection (ID) and slot filling (SF) tasks is divided into two phases: an implicit interaction phase and an explicit interaction phase. In the implicit interaction phase, ID and SF tasks transmit information through the bidirectional recurrent neural network, and in the explicit interaction phase, ID and SF tasks fuse information through a collaborative mechanism. This interaction achieved the most advanced results at the time. In 2021, Qin et al. [28] used a co-interactive module to establish a bidirectional connection between two tasks. This module allows slots and intents to pay attention to each other’s information, to better capture the cross-influence between the two tasks, and this interaction is evaluated and improved on ATIS and SNIPS. In 2022, Qin et al. [12] added the correlation between the learning intents and slots of three contrast learning modules by introducing the contrastive learning framework. The results of cross-lingual intent detection and slot filling have been significantly increased on the MultiATIS++ dataset. It is therefore promising to consider the interaction between these two tasks.
To sum up, strategies based on multilanguage models are commonly used in the field of cross-lingual transfer learning. The advantages of such strategies are that they can reduce training costs, improve efficiency, and enhance generalization ability. In addition, the bidirectional interactive framework is often used in the joint learning framework, which can make the model capture more shared knowledge between tasks, thus improving the performance of both tasks. Second, the bidirectional interactive framework helps to improve the interpretation ability of the model, making it easier to analyze the effects between slots and intents.
Therefore, this paper will adopt the strategy of a multilanguage model and propose an attention-based bidirectional interactive framework based on a bidirectional interactive framework to realize cross-lingual intent detection and slot filling in the field of in-vehicle dialogue.

3. Modeling Framework

The proposed overall framework shown in Figure 1 has two parts: the encoder based on contrastive learning and the decoder based on interactive attention. The encoder is mainly responsible for converting text characters into embedded vectors for deep learning and projecting the vectors onto a high-dimensional plane. Contrastive learning can help models learn discrimination feature representations. By comparing the source statement with the positive and negative samples, the model learns the feature representation that can accurately distinguish the positive and negative samples. The positive and negative samples in this paper are multilingual text representations with the same semantics and different semantics as the source text, respectively. Therefore, the encoder based on contrast learning uses high-dimensional embedded vectors to carry out standardized multilanguage learning through self-supervised means, which enables the encoder to integrate more language information in the coding stage and increase the cross-lingual ability of the model. The decoder based on interactive attention uses the characteristics of coding to generalize intent detection and slot filling. In this process, these features are extracted and used for learning, while interactive attention is used for two-way interaction between intent detection and slot filling.

3.1. Contrastive-Learning-Based Encoder

An encoder converts an input sequence into a fixed-length embedding, such as a piece of text. This embedding is called a “context vector” or “encode vector”. The code vector contains information about the input sequence and acts as the input to the decoder to help it generate the output sequence.
Contrastive learning (CL) is a machine learning method that learns differences by comparing positive and negative samples. Contrastive learning helps NLP models learn meaningful representations of text data. The model learns to capture semantic and contextual information by contrasting positive and negative samples, enabling better understanding of word meanings, sentence structures, and overall language semantics. It has been widely used in computer vision, natural language processing, and other fields, and performs well [29].
As shown in Figure 2, the model mainly consists of a multilanguage pretraining model and a contrastive learning module. The multilanguage pretraining model is pretrained based on various texts, which can be mBERT, XLM-R, and other models. Contrastive learning modules include a sentence-level intent CL module, word-level slot CL module, and semantic level intent-slot CL module. The sentence-level intent CL module is used to distinguish the intents of texts in different languages with the same semantics; the word-level slot CL module is used to distinguish the slots of words in different languages with the same semantics; and semantic-level intent-slot CL module is used to learn the correlation between intent slots.
The contrastive learning adopted in this paper is to take the new sentence expression obtained from the translation of the words in the source sentence as the positive sample and the expression with different meanings as the negative sample.

3.1.1. Positive Sample

A positive sample is a multilingual statement with the same semantics. First, given a source statement X = x 1 , x 2 , x 3 , , x n , a positive sample X + = x 1 + , x 2 + , x 3 + , , x n + is generated by a bilingual dictionary. Specifically, a positive sample is obtained by randomly selecting each word x 1 in X and replacing x 1 with a translation improved by a bilingual dictionary. For example, given the English source statement watch football match, the positive sample is 看 (zh/watch) サッカー (ja/football) Спички (es/match).

3.1.2. Negative Sample

A negative sample is very different from a positive sample. A source statement generally generates only one positive sample statement, but a negative sample is a collection of statements. The negative sample set is generated in such a way that the coding samples of previous batches will be reused, thus eliminating unnecessary negative coding processes.
The whole encoder process: First of all, the source utterance X carries out multilanguage code switch data to obtain X +   and X   , and then input to the mBERT encoder to obtain the representation H , H + , H .

3.1.3. Loss Function of Contrastive Learning

Three contrastive learning modules consisting of positive samples X +   , negative samples X   , and source utterances X as inputs are the sentence-level intent CL module, the word-level slot CL module, and the semantic-level intent-slot CL module.
  • Sentence-level intent CL module for aligning sentence representations across languages for intent detection.
Intent detection is a sentence-level classification task. In the cross-linguistic domain, aligning sentence representations is an important measure for the zero-shot cross-linguistic intent detection task. Sentence alignment can explicitly encourage models to align similar sentence representations into the same space across languages for intent detection. Here is the sentence-level intent contrastive learning loss calculated for a sentence:
L U I = log s ( h C L S , h C L S + ) s ( h C L S , h C L S + ) + s ( h C L S , h C L S )
2.
Word-level slot CL module for cross-lingual alignment of slot representations for slot filling.
Since slot filling is a word-level task, cross-lingual alignment of sentence representations is used to help the model consider the alignment of slot filling for fine-grained cross-lingual transfer. Here is the word-level slot contrastive learning loss for the i-th word:
L T S i = i = 1 n log s ( h i , h i + ) s ( h i , h i + ) + s ( h i , h i ) / n
3.
Semantic-level intent-slot CL module for aligning representations between slots and intents.
When slot and intent belong to the same object, they are usually highly semantically related. Therefore, this paper argues that intents and their slots in sentences constitute a positive correlation, while the corresponding slots in other sentences constitute a negative correlation. Therefore, this paper further introduces semantic-level intent-slot CL loss to model the semantic interaction between slots and intents, which may further improve the cross-linguistic transfer between them:
L S I S 1 = i = 1 n log s ( h C L S + , h i ) s ( h C L S + , h i ) + k = 0 K 1 s ( h C L S , h i , k ) / n
L S I S 2 = i = 1 n log s ( h C L S , h i + ) s ( h C L S , h i + ) + k = 0 K 1 s ( h C L S , h i , k ) / n
L S I S = L S I S 1 + L S I S 2
where L S I S 1 and L S I S 2 are the contrastive learning loss between the positive samples of the intent vector and slot vector and the source utterance, respectively, and s ( a , b ) represents the dot product of a and b .

3.2. Interactive-Learning-Based Decoder

In neural networks, decoders play a crucial role in various tasks that involve generating output from abstract representations or latent features. Decoders are often paired with encoders, which are responsible for converting input data into a compressed or abstract representation. The decoder’s role is to transform these representations back into the desired output format.
The decoder stage adds an interactively learned multitask attention head to the fully connected layer. Because of the high correlation between the intent detection and slot filling tasks, the hidden state of the slot filling task is shared to the intent detection task through the attention mechanism in the decoder phase, and the hidden state of the intent detection task acts on the slot filling as well. Intention detection task and slot filling task learn together and promote each other.
The intent detection attention head and slot filling attention head are computed based on the output sequence of the pretrained model by calculating the attention computation score, which is combined with the fully connected layer for different tasks to compute the final state.
  • Intent detection task attention head:
α I = e x p h C L S e x p h C L S + i = 1 n exp h i
2.
Slot filling task attention header:
α i S = e x p h i e x p h C L S + i = 1 n exp h i
The intent detection task is to put h C L S into the classification layer to find labels:
o I = s o f t m a x   W I h C L S o u t + α i S h i o u t + b I
There, W and b are the matrix of the parameter to be trained. In this paper, we jointly fine-tune all the parameters of mBERT and W by maximizing the logarithmic probability of correct labeling.
For the sequence labeling task, this paper feeds the final hidden state of the model to the softmax layer to classify the slots. In this paper, the hidden state corresponding to the first sublabel is used as input to classify words:
o S = s o f t m a x W S   h i o u t + α I h C L S o u t + b S , i = 1 , 2 , n

3.3. Multitask Learning

The loss function is to calculate the difference between the forward computed result and the true value of each iteration of the neural network, to guide the next step of training in the right direction. Its input is the network prediction and the real target value, and then calculates a distance value, to measure the quality of the network output results. The loss function consists of the loss of the contrastive learning module ( L U I , L T S i , L S I S ) and the loss of the multitasking ( L I , L S ), and the different structures of the loss function are connected by the hyperparameter λ .
  • Loss of intent detection task:
L I i = 1 n I y ^ i I log o i I
2.
Loss of slot filling task:
L S j = 1 n i = 1 n S y ^ j i , S log o j i , S
The total loss function is a linear combination of the losses from the three comparison learning modules and the two tasks:
L = λ I L I + λ S L S + λ L I L L I + λ L S L L S + λ G I S L G I S

4. Experimental Setup

4.1. Experimental Data

In this paper, we conduct experiments on the public datasets MultiATIS++ and the in-vehicle dialogue datasets (MIvD) collected by our own constructed in-vehicle dialogue platform in the vehicle domain. MultiATIS++ contains nine languages, including English (en), Spanish (es), Portuguese (pt), German (de), French (fr), Chinese (zh), Japanese (ja), Hindi (hi), and Turkish (tr). The detailed item information corresponding to the datasets is given in Table 1.
MIvD is collected by the onboard device and then manually redacted and annotated. It includes annotations of intent and semantic slots in four main languages, namely, English (en), Chinese (zh), Japanese (ja), and Arabic (ar). The details are summarized in the following Table 2.
To describe the dataset in more detail, this paper is carried out in Table 3 to show some features of the dataset. Utterance is the instruction entered by the user, and increases the robustness of the model; we have not removed the colloquial words. Intent types are the intent labels of the input instructions, and slot values are the ranges of the attributes of the intent labels, and some of the instructions’ slot values are empty.

4.2. Evaluation Indicators and Baseline Model

4.2.1. Evaluation Indicators

In this paper, there are three evaluation indicators used in this paper: the Intent Accuracy for the intent detection task, slot F1 for the slot filling task, and the combined indicators overall accuracy for the two tasks. These evaluation indicators are also widely used in other domains of NLU.
  • Intent accuracy: Intent Accuracy is used to evaluate the performance of intent detection by calculating the percentage of sentences that correctly predict the intent.
  • Slot F1: The performance of slot filling is evaluated using the F1 score, which is the average score of the reconciliation between accuracy and recall. Slot predictions are considered correct when exact matches are found [30].
  • Overall accuracy: It calculates the proportion of sentences that correctly predict intents and slots. The indicator takes into account both intent detection and slot filling [13].

4.2.2. Baseline Model

  • CoSDA-ML. Qin et al. [12] proposed a dynamic code-switching method for randomly performing multilingual word-level substitutions. For a fair comparison of MIvD, this paper uses Chinese training data and code-switching data for fine-tuning.
  • Multilingual-ZeroShot. Train a joint intent detection and slot filling model using English and generalize to other languages by Jitin [31].
  • GL-CLEF is a global–local contrastive learning framework for display alignment proposed by Qin et al. [13].
  • LAJ-MCJ is a multilevel label-aware contrastive learning framework for display alignment that was proposed by Liang et al. [32].

4.3. Experimental Parameters

In this paper, we use the multilingual mBERT model. The mBERT model has the same architecture and training process as the BERT model, which has 12 transformer layers, 768 hidden states, and 12 attention heads. As shown in Table 4, the optimal hyperparameters are selected by comparing the combination of batch size and learning rate. The probability of replacing a word in the contrastive learning module during training is 0.55. For fine-tuning, the maximum length of the sequence is 128, and the batch size is 32. The optimizer chooses Adam with an initial learning rate of 5 × 10−6 and a dropout probability of 0.1. The range of the learning rate is as follows: {2 × 10−7, 5 × 10−7, 1 × 10−6, 2 × 10−6, 5 × 10−6, 6 × 10−6, 5 × 10−5, 5 × 10−4}; batch size {4, 8, 16, 32}; the size of negative samples chosen from {4, 8, 16, 32}, which are the weight parameters of sentence-level intent CL modules, word-level slot CL modules, and semantic-level intent-slot CL modules, respectively.
Hardware labs are equipped with consumer-grade RTX 3090, enterprise-grade A40 general-purpose AI servers, and enterprise-grade high-performance AI training server A100, with the total GPU computing power FP64 ≥ 200 TFlops and FP32 ≥ 3.0 PFlops. The software uses Linux-based miniconda3 with pytorch1.12.1 + python3.7.

5. Result Analysis and Discussion

The IABCL framework proposed in this paper has attained superb results on both MultiATIS++ and MIvD datasets, where the results are excellent on the MIvD dataset. To evaluate the effectiveness of the IABCL-based framework, this paper investigates the following three questions.
  • RQ1: Does IABCL predict better results than other cross-lingual migration models in the MIvD dataset?
  • RQ2: How well does IABCL work on public datasets?
  • RQ3: What role do contrastive learning and interactive attention specifically play?
RQ1: Does IABCL predict better results than other cross-lingual migration models in the MIvD dataset?
To verify this problem, this paper conducts a comparative experiment for the IABCL model on MIvD with other basic models, and the results of the experiment are shown in Table 5.
For the experiments on MIvD, this paper uses mBERT. As shown in Table 5, this paper compares the performance of four different models on different languages and metrics, including intent accuracy, slot F1 value, and overall accuracy. Intent, slot, and overall results have the same trend in different languages. As shown in Figure 3 (results of overall accuracy), IABCL obtained optimal results on MIvD. Compared with other models, this paper improves more significantly in Japanese and English. Because of the cost involved, Multilingual-ZeroShot needs to translate each dataset, and this experiment only translates one-third of the vocabulary comparable with other split-word translations.
Since IABCL and other baseline models are trained based on the Chinese language, they have excellent performance in Chinese onboard datasets; e.g., IABCL achieves 97.40% in Chinese intent accuracy, 95.72% in slot f1 value, and 94.07% in overall accuracy. It is worth noting that in cross-lingual transfer learning, the results of Japanese are better than those of English and Arabic because Chinese and Japanese are of similar language families, and the results of English and Arabic are slightly inferior to those of Japanese, which further verifies that the effect of cross-lingual transfer learning in the same language family is more effective than that of cross-lingual transfer between languages of different language families. Meanwhile, we also analyze the results in English and Arabic separately. The MIvD dataset contains high-dimensional semantic data, such as “I am sweating”, “Here we go”, and “كثيراً حلوة الأغنية هذه” (This song is very sweet), which requires more complex semantic modeling to understand. Additionally, Arabic is a right-to-left language, and this style also affects the final result. Relative to the three baseline models, the framework proposed in this paper excels in cross-lingual transfer learning ability, comparing CoSDA-ML, Multilingual-ZeroShot, and GL-CLEF, with an average improvement of 5.63%, 25.07%, and 3.27% in intent accuracy in English, Japanese, and Arabic, respectively, and an improvement of 4.59%, 24.63%, 2.24%, and 3.27% in slot F1 values, respectively, and 24.63%, 2.24%, and 5.83%, 35.79%, and 0.54% in overall accuracy, respectively.
Overall, the adoption of the IABCL framework has yielded excellent results in both intent and slot prediction in the field of in-vehicle dialogue system.
RQ2: How well does IABCL work on public datasets?
In this paper, the IABCL framework is validated on the public datasets MultiATIS++, and to validate the impact of different multilingual pretraining models on the performance of this model, two pretraining models, mBERT and XLM-R, are used for comparison, and the experimental results are shown in Table 6. The IABCL framework with mBERT achieves the optimal intent accuracy compared with other models, and the F1 value and overall accuracy are slightly worse, while the IABCL framework with XLM-R achieves the optimal slot F1 and overall accuracy, and is slightly worse in intent accuracy, but also achieves more than 90% in general.
As shown in Figure 4, comparing mBERT (2019) and XLM-R (2020) models, this paper improves by 16.42%, 14.75% and 18.61%, 16.01% in slot F1 and overall accuracy, respectively, which verifies that code switch improves significantly on cross-lingual migration. Compared with CoSDA-ML, this paper improves by 1%, 7%, and 6% in intent, slot, and overall, verifying that contrastive learning works significantly on cross-lingual transfer learning. Compared with LAJ-MCJ, this paper only improves the result of intent on mBERT, and the slot and overall are reduced. According to our analysis, in the generation of positive and negative samples of contrastive learning, to reduce the cost, this paper simply takes other utterances from the same batch as negative samples, which may result in the existence of “false-negative samples”, and the final accuracy will be reduced; at the same time, in XLM-R, the accuracy is reduced by 1%, 7%, and 6%. This may cause the existence of “false-negative samples”, and the final accuracy will be reduced; at the same time, in the XLM-R model, the results of slots and overall have been improved, and the intent has been reduced, which verifies that the existence of the “false-negative samples“ will affect the final results.
Taken together, the IABCL model performs better in cross-lingual intent detection and slot filling tasks, and the performance of the IABCL model needs to be improved in Arabic and English. Future research can focus on improving the model’s ability to transfer across language families and exploring more effective methods to achieve cross-lingual transfer learning.
RQ3: What are the specific roles of contrastive learning and interactive attention?
To answer the question, in this paper, we remove the proposed method modules one by one and perform ablation experiments in the MIvD dataset. The results are shown in Table 7. In the table, our-cl is the removal of the contrast learning module, our-bi_interaction is the removal of the attention interaction module, our-all is the removal of both contrast learning and attention interaction modules, and our is the complete model architecture.
Comparing our-all and our-cl (i.e., containing only interactive attention) improves the results in intent, slot, and overall, so the joint framework enables the model to learn more implicit information. Additionally, comparing our and our-cl improves the effect in Sino-Tibetan languages (ZH, JA) more than our-bi_interaction, and we conclude that interactive attention can better act on cognate languages. Further, according to our-bi_interaction, our-all, and our, the boosting effect of contrastive learning in non-Chinese-Tibetan languages is obvious, and contrastive learning can promote cross-linguistic learning of the model and improve the compatibility of the model.
Therefore, in summary, both interactive attention and contrastive learning have their advantages, and the combination of the two is more instructive and practicable.

6. Conclusions

To improve the cross-lingual intent detection and slot filling ability of the in-vehicle dialogue system, this paper proposes an interactive attention-based contrastive learning framework (IABCL). Meanwhile, for the first time, this paper constructs multilingual datasets (MIvD) in the field of in-vehicle dialogue, then uses the IABCL framework to verify it, and finally yielded superior outcomes. The experimental results show that the transfer learning can also be centered on the Chinese language and transfer to other languages to obtain a comparable effect.
However, the current results in English and Arabic need to be improved, and we have found that there is a certain correlation between different languages of the same language family, and that the universality of cross-language transfer learning for one or more languages needs to be improved. In the future, based on the Chinese language family, we will further improve the cross-language transfer learning ability of the proposed method to improve the generalization ability of the model.

Author Contributions

Conceptualization, L.F. and X.L.; methodology, L.F. and X.L.; software, L.F.; validation, L.F. and P.C.; formal analysis, L.F. and X.L.; investigation, L.F. and L.Z.; resources, L.F. and X.L.; data curation, L.F.; writing—original draft preparation, L.F.; writing—review and editing, L.F., X.L., and P.C.; visualization, L.F. and L.Z.; supervision, P.C.; project administration, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by NSFC under Grant No. 6276085 and the Graduate Innovation Project of Hefei University under Grant No. 21YCXL20 and the 2023 Anhui University Research Project of China No. 2023AH052179.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing does not apply to this article.

Acknowledgments

The authors gratefully acknowledge the National Intelligent Voice Innovation Center for their wonderful support during the process of establishing the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hall, T.; Beecham, S.; Bowes, D.; Gray, D.; Counsell, S. A systematic literature review on fault prediction performance in software engineering. IEEE Trans. Softw. Eng. 2011, 38, 1276–1304. [Google Scholar] [CrossRef]
  2. Liu, S.; Qi, Z.; Wang, Y.; Tang, Q.; He, Y.; Liu, Y. Design and analysis of dialogue based on vehicle. Veh. Electr. Appl. 2023, 1, 46–48. (In Chinese) [Google Scholar] [CrossRef]
  3. Zhang, L. Research and Application of Vehicle-Mounted Human-Computer Interface Interaction Design for Voice Interaction; Beijing University of Posts and Telecommunications: Beijing, China, 2021. (In Chinese) [Google Scholar]
  4. Du, Z. Research on dialogue System of Vehicle HUD Based on Gesture Recognition; Hebei University of Science and Technology: Shijiazhuang, China, 2019. (In Chinese) [Google Scholar]
  5. Gao, Q.; Xu, Y. Dialogue system of vehicle touch control. Instrum. Technol. Sens. 2012, 2, 72–74. (In Chinese) [Google Scholar]
  6. Khalil, T. Cross-lingual intent classification in a low resource industrial setting. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019. [Google Scholar]
  7. Arora, A. Cross-lingual transfer learning for intent detection of COVID-19 utterances. In Proceedings of the Workshop on NLP for COVID-19 (Part 2) at EMNLP 2020, Online, 15 August–4 September 2020. [Google Scholar]
  8. Gerz, D.; Su, P.-H.; Kusztos, R.; Mondal, A.; Lis, M.; Singhal, E.; Mrkšić, N.; Wen, T.-H.; Vulić, I. Multilingual and Cross-Lingual Intent Detection from Spoken Data. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; Association for Computational Linguistics: Punta Cana, Dominican Republic, 2021; pp. 7468–7475. [Google Scholar]
  9. Liu, Z.; Winata, G.I.; Lin, Z.; Xu, P.; Fung, P. Attention-informed mixed-language training for zero-shot cross-lingual task-oriented dialogue systems. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8433–8440. [Google Scholar]
  10. Xu, W.; Haider, B.; Mansour, S. End-to-End Slot Alignment and Recognition for Cross-Lingual NLU. arXiv 2020, arXiv:2004.14353. [Google Scholar]
  11. FitzGerald, J.G. STIL-Simultaneous Slot Filling, Translation, Intent Classification, and Language Identification: Initial Results using mBART on MultiATIS++. arXiv 2020, arXiv:2010.00760. [Google Scholar]
  12. Qin, L. Cosda-ml: Multi-lingual code-switching data augmentation for zero-shot cross-lingual nlp. arXiv 2020, arXiv:2006.06402. [Google Scholar]
  13. Qin, L.; Chen, Q.; Xie, T. GL-CLeF: A Global-Local Contrastive Learning Framework for Cross-lingual Spoken Language Understanding. arXiv 2022, arXiv:2204.08325. [Google Scholar]
  14. Yang, B. Design and Implementation of Vehicle Speech Control System Based on Oral Comprehension by Rules and Statistical Methods; Hebei University of Science and Technology: Shijiazhuang, China, 2018. (In Chinese) [Google Scholar]
  15. Rafiepour, M.; Sartakhti, J.S. CTRAN: CNN-Transformer-based Network for NLU. arXiv 2023, arXiv:2303.10606. [Google Scholar]
  16. Ruder, S.; Vulić, I.; Søgaard, A. A survey of cross-lingual word embedding models. J. Artif. Intell. Res. 2019, 65, 569–631. [Google Scholar] [CrossRef]
  17. Zhuang, F.; Luo, P.; He, Q.; Shi, Z. Research progress of transfer learning. J. Softw. Sci. 2014, 26, 26–39. (In Chinese) [Google Scholar]
  18. Lefevre, F.; Mairesse, F.; Young, S. Cross-lingual spoken language understanding from unaligned data using discriminative classification models and machine translation. In Proceedings of the Eleventh Annual Conference of the International Speech Communication Association, Chiba, Japan, 26–30 September 2010. [Google Scholar]
  19. Adams, O. Cross-lingual word embeddings for low-resource language modeling. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, Valencia, Spain, 3–7 April 2017. [Google Scholar]
  20. Johnson, A. Cross-lingual transfer learning for Japanese named entity recognition. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Industry Papers), Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar]
  21. Li, Y. Cross-Lingual Multi-Task Neural Architecture for Spoken Language Understanding. In Proceedings of the Interspeech, Hyderabad, India, 2–6 September 2018. [Google Scholar]
  22. Schuster, S.; Gupta, S.; Shah, R.; Lewis, M. Cross-lingual transfer learning for multilingual task-oriented dialog. Proceedings of Annual Conference of the North American Chapter of the Association for Computational Linguistics, Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar]
  23. Zhang, X.; Wang, H. A joint model of intent determination slot filling for spoken language understanding. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016. [Google Scholar]
  24. Wang, Y.; Shen, Y.; Jin, H. A bi-model based on semantic frame parsing model for intent detection and slot filling. arXiv 2018, arXiv:1812.10235. [Google Scholar]
  25. Liu, B.; Lane, I. Attention-based recurrent neural network models for joint intent detection and slot filling. In Proceedings of the Interspeech, San Francisco, CA, USA, 8–12 September 2016. [Google Scholar]
  26. Chen, Q.; Zhuo, Z.; Wang, W. Bert for joint intent classification and slot filling. arXiv 2019, arXiv:1902.10909. [Google Scholar]
  27. Zhou, P.; Huang, Z.; Liu, F.; Zou, Y. PIN: A novel parallel interactive network for spoken language understanding. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021. [Google Scholar]
  28. Qin, L. A co-interactive transformer for joint slot filling and intent detection. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021. [Google Scholar]
  29. Zhang, C.; Chen, J.; Li, Q.; Deng, B.; Wang, J.; Chen, C. Summary of deep comparative learning. Acta Autom. Sin. 2023, 49, 15–39. (In Chinese) [Google Scholar] [CrossRef]
  30. Qin, L. A survey on spoken language understanding: Recent advances and new frontiers. arXiv 2021, arXiv:2103.03095. [Google Scholar]
  31. Krishnan, J.; Anastasopoulos, A.; Purohit, H.; Rangwala, H. Multilingual Code-Switching for Zero-Shot Cross-Lingual Intent Prediction and Slot Filling. In Proceedings of the 1st Workshop on Multilingual Representation Learning, Online, 11 November 2021; Association for Computational Linguistics: Punta Cana, Dominican Republic, 2021; pp. 211–223. [Google Scholar]
  32. Liang, S. Label-aware Multi-level Contrastive Learning for Cross-lingual Spoken Language Understanding. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022. [Google Scholar]
Figure 1. Interactive attention model based on contrastive learning.
Figure 1. Interactive attention model based on contrastive learning.
Sensors 23 08501 g001
Figure 2. Contrastive-learning-based encoder.
Figure 2. Contrastive-learning-based encoder.
Sensors 23 08501 g002
Figure 3. Overall accuracy of MIvD in different languages under baseline.
Figure 3. Overall accuracy of MIvD in different languages under baseline.
Sensors 23 08501 g003
Figure 4. Contrastive learning-based encoder. (a) Experimental results of different baseline models under mBERT. (b) Experimental results of different baseline models under XLM-R.
Figure 4. Contrastive learning-based encoder. (a) Experimental results of different baseline models under mBERT. (b) Experimental results of different baseline models under XLM-R.
Sensors 23 08501 g004
Table 1. MultiATIS++ dataset.
Table 1. MultiATIS++ dataset.
LanguageUtterancesIntent TypesSlot Types
TrainDevTest
en44884908931884
es44884908931884
pt44884908931884
de44884908931884
fr44884908931884
zh44884908931884
ja44884908931884
hi14401608931775
tr578607151771
Table 2. MIvD dataset.
Table 2. MIvD dataset.
LanguageUtterancesIntent NumberSlot Number
TrainDevTest
zh22,152277027691714
ja-3923921714
en-5015001714
ar-4934931714
Table 3. MIvD dataset description.
Table 3. MIvD dataset description.
LanguageUtteranceIntent TypesSlot Values
zh哎帮我把空调调一下19度adjust_ac_temperture19
ja風量を九にadjust_ac_windspeed
enOpen ac cooling modeopen_ac_modecooling
arدرجة ثلاثين على الحرارة درجة تعيين.adjust_ac_temperature ثلاثين
Table 4. Experiment-related parameters.
Table 4. Experiment-related parameters.
Experimental Modules and ParametersParameters and Module-Specific Information
Pretrained modelsmBERT (12 transformer layers, 768 hidden layers, and 12 attention heads)
Batch size32
Learning rate5 × 10−6
Probability of code switching0.55
Maximum sequence length128
OptimizerAdam
Dropout0.1
Number of negative samples16
λ10.01
λ20.005
λ30.01
Table 5. Performance comparison on MIvD.
Table 5. Performance comparison on MIvD.
Intent Accuracy (%)ZHJAARENAvg
CoSDA-ML (2020)96.2460.7160.3260.3369.40
Multilingual-ZeroShot (2021)96.7861.2230.2231.6054.95
GL-CLEF (2022)97.3272.1961.8654.471.44
IABCL97.4073.9762.6761.6073.86
Slot F1 (%)ZHJAARENavg
CoSDA-ML (2020)93.0072.2726.7753.5461.39
Multilingual-ZeroShot (2021)90.4236.5219.4736.4845.72
GL-CLEF (2022)95.5375.5428.6255.4863.79
IABCL95.7279.6830.5456.1565.22
Overall Acc (%)ZHJAARENavg
CoSDA-ML (2020)90.0846.6828.0043.5552.07
Multilingual-ZeroShot (2021)88.7615.005.477.9029.28
GL-CLEF (2022)93.7158.9228.837.2054.66
IABCL94.0760.4528.8946.457.33
Table 6. Performance comparison on MultiATIS++. Results with * are from LAJ-MCJ.
Table 6. Performance comparison on MultiATIS++. Results with * are from LAJ-MCJ.
MethodmBERTXLM-R
Intent Acc (%)Slot F1 (%)Overall Acc (%)Intent Acc (%)Slot F1 (%)Overall Acc (%)
mBERT (2019) *88.4261.6636.29---
XLM-R (2020) *---93.0257.3833.31
CoSDA-ML (2020)90.8768.0843.1593.0470.0143.72
Ensemble-Net (2021) *87.2055.78----
LAJ-MCJ (2022)92.4178.2352.5093.4975.6947.58
IABCL92.9678.0851.0493.1375.9949.38
Table 7. Ablation experiments.
Table 7. Ablation experiments.
Intent Acc (%)ZHJAARENAvg
Our-cl97.3273.2760.9858.9972.64
Our-bi_interaction96.9373.1961.8660.0072.99
Our-all95.6672.9559.2257.8071.40
Our97.4073.9762.6761.6073.91
Slot F1 (%)ZHJAARENAvg
Our-cl95.5376.8229.8455.0062.29
Our-bi_interaction95.3575.5429.9255.4864.07
Our-all93.6375.3028.6254.1762.93
Our95.7279.6830.5456.1565.52
Overall Acc (%)ZHJAARENAvg
Our-cl93.7159.8228.8041.4055.93
Our-bi_interaction93.6459.1828.8943.8056.37
Our-all92.6655.9527.1839.8053.89
Our94.0760.4530.2246.4057.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Fang, L.; Zhang, L.; Cao, P. An Interactive Framework of Cross-Lingual NLU for In-Vehicle Dialogue. Sensors 2023, 23, 8501. https://doi.org/10.3390/s23208501

AMA Style

Li X, Fang L, Zhang L, Cao P. An Interactive Framework of Cross-Lingual NLU for In-Vehicle Dialogue. Sensors. 2023; 23(20):8501. https://doi.org/10.3390/s23208501

Chicago/Turabian Style

Li, Xinlu, Liangkuan Fang, Lexuan Zhang, and Pei Cao. 2023. "An Interactive Framework of Cross-Lingual NLU for In-Vehicle Dialogue" Sensors 23, no. 20: 8501. https://doi.org/10.3390/s23208501

APA Style

Li, X., Fang, L., Zhang, L., & Cao, P. (2023). An Interactive Framework of Cross-Lingual NLU for In-Vehicle Dialogue. Sensors, 23(20), 8501. https://doi.org/10.3390/s23208501

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop