Computational Intelligence and Neuroscience

Computational Intelligence and Neuroscience / 2021 / Article
Special Issue

Interpretation of Machine Learning: Prediction, Representation, Modeling, and Visualization 2021

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 9975078 |

Chenggang Mi, Shaolin Zhu, Rui Nie, "Improving Loanword Identification in Low-Resource Language with Data Augmentation and Multiple Feature Fusion", Computational Intelligence and Neuroscience, vol. 2021, Article ID 9975078, 9 pages, 2021.

Improving Loanword Identification in Low-Resource Language with Data Augmentation and Multiple Feature Fusion

Academic Editor: Nian Zhang
Received09 Mar 2021
Revised18 Mar 2021
Accepted25 Mar 2021
Published08 Apr 2021


Loanword identification is studied in recent years to alleviate data sparseness in several natural language processing (NLP) tasks, such as machine translation, cross-lingual information retrieval, and so on. However, recent studies on this topic usually put efforts on high-resource languages (such as Chinese, English, and Russian); for low-resource languages, such as Uyghur and Mongolian, due to the limitation of resources and lack of annotated data, loanword identification on these languages tends to have lower performance. To overcome this problem, we first propose a lexical constraint-based data augmentation method to generate training data for low-resource language loanword identification; then, a loanword identification model based on a log-linear RNN is introduced to improve the performance of low-resource loanword identification by incorporating features such as word-level embeddings, character-level embeddings, pronunciation similarity, and part-of-speech (POS) into one model. Experimental results on loanword identification in Uyghur (in this study, we mainly focus on Arabic, Chinese, Russian, and Turkish loanwords in Uyghur) showed that our proposed method achieves best performance compared with several strong baseline systems.

1. Introduction

Bilingual data play an very important role in cross-lingual natural language processing (NLP) tasks, such as cross-lingual text classification, cross-lingual information retrieval, and neural machine translation. However, bilingual data are often difficult to obtain. Lexical borrowing happens in almost every language; Figure 1 gives several loanwords in Uyghur (the reasons why we choose Uyghur as an example in our study are as follows: (1) there are many loanwords in Uyghur and (2) Uyghur is a low-resource language). If loanwords in low-resource languages can be identified effectively, it will be a novel way to alleviate the data sparseness existing in many cross-lingual NLP tasks.

Loanword identification is a task of finding out loanwords of a specific language (donor language) in texts in another language (receipt language). There are about three kinds of loanword identification methods: (1) rule-based method; (2) statistical-based method; and (3) deep learning-based method. Early studies on loanword identification often based on rules. For example, McCoy and Frank [1] proposed a string similarity-based loanword identification model that relies on the ED algorithm. With the development of machine learning algorithms in NLP area, statistical-based methods are also proposed [2]. In recent years, deep learning algorithm such as bidirectional LSTM and convolutional neural network (BLSTM + CNN) are also used in loanword identification tasks [3]. Due to the lack of generalization ability of rule-based methods and limitation of training data in statistical-based methods, recent studies often combine the rule and statistical features together to improve the model performance effectively [4, 5]. However, almost all of these methods suffer from data sparseness during model training, especially in low-resource settings.

As a common used method to alleviate the data sparseness, data augmentation is one of the most popular methods in this topic. For example, Liu et al. [6] proposed to use a GAN model consisting of two generators and one discriminator to produce meaningful natural language sentences. Motivated by this study, we propose to use a lexical constraint-based data augmentation model to generate more training data for loanword identification. Different from [6], we take the loanwords in training data as a lexical constraint to produce more sentences containing the loanwords.

After investigation, we find that there are two important clues in loanword identification: semantic similarity and pronunciation similarity. To incorporate these two features into one feature, we propose to transfer the semantic similarity as word-level feature and pronunciation similarity as character-level feature. Then, we fuse these two features into one feature. Meanwhile, we incorporate the fusion feature, pronunciation feature, and POS feature into a log-linear RNN to achieve the best performance in loanword identification.

The main contributions of this study are as follows:(i)First, a lexical constraint-based data augmentation method is proposed to generate more training data for loanword identification task.(ii)Second, we incorporate multilevel features, pronunciation similarity feature, and POS feature into a log-linear RNN model to improve the performance of the loanword identification model for low-resource language.(iii)Third, we conduct an experiment on loanword (Arabic, Chinese, Russian, and Turkish) identification in Uyghur; experimental results show that our proposed model achieves the best performance compared with several strong baseline systems.

The rest of this paper is organized as follows. Section 2 introduces some recent studies related to our topic. We present details of our proposed method in Section 3. Datasets, settings, and experimental results are described in Section 4. We show the analysis of experimental results in Section 5. In Section 6, we conclude this study and give some possible future directions.

In this section, we present some work related to our study.

2.1. Loanword Identification

Lexical borrowing has received relatively little attention in natural language processing area. Tsvetkov and Dyer [7] proposed a morph-phonological transformation model to obtain good performance at predicting donor forms from borrowed forms. Tsvetkov et al. [7] suggested to use the lexical borrowing as a model in an SMT framework to translate OOV words. Gerz et al. [8] analyzed the implication of variation in structural and semantic properties in general language-independent architectures on the language modeling task. Mi et al. [9] used shallow features such as string similarity to detect loanwords in Uyghur. Mi et al. [3] presented a neural network-based loanword identification model that also incorporated several shallow features. However, these methods only trained loanword identification models based on some monolingual corpora. It fails to project donor language and receipt language into one semantic space. The limitation of training data also exists.

2.2. Data Augmentation for NLP

The main goal of data augmentation in NLP is to generate additional, synthetic data using the data you have to alleviate the data sparseness during model training [10]. There are several data augmentation methods in NLP area [11]. The first one is lexical substitution which tries to substitute words present in a text without changing the meaning of the sentence [12]. The second one is back translation, which is commonly used in neural machine translation (NMT). Back translation first trains an intermediate system on the parallel data which is used to translate the target monolingual data into the source language. The result is a parallel corpus where the source side is synthetic machine translation output while the target is genuine text written by humans. The synthetic parallel corpus is then simply added to the real bitext in order to train a final system that will translate from the source to the target language [13]. The syntax-tree manipulation has been used in [14]; the idea is to parse and generate the dependency tree of the original sentence, transform it using rules, and generate a paraphrased sentence. Mixup is a simple yet effective image augmentation technique introduced by Zhang et al. [15]. The idea is to combine two random images in a mini-batch in some proportion to generate synthetic examples for training. The most recent data augmentation method is generative model; this kind of method tries to generate additional training data while preserving the class label [16].

2.3. Sequence Labeling in NLP

There are two main types of sequence labeling methods in NLP, such as gradient-based methods and search-based methods [17]. As for the probabilistic gradient-based learning methods such as conditional random fields (CRFs) and recurrent neural network (RNN), they have high accuracy because of the exact computation of the gradient and probabilistic information. Nevertheless, those methods have critical drawbacks. First, the probabilistic gradient-based methods typically do not support search-based optimization (search-based learning or decoding-based learning), which is important in sequence labeling problems with emphasis on the learning speed (e.g., for large-scale datasets). In tasks with complex structures, gradient computation is usually quite complicated sometimes and even intractable. This is mainly because dynamic programming for computing gradient is hard to scale for large-scale datasets. On the other hand, the search technique is easier to scale to large-scale datasets. This is because search-based learning is much simpler than gradient-based learning [1820]—just search the promising output candidates and compare them with the oracle labels and update the weights accordingly. Another category of sequence labeling methods is the search-based learning methods (i.e., decoding-based learning), such as structured perceptron and MIRA. A major advantage of those methods is that they support search-based learning, such that the gradient is not needed and the learning is done by simply searching and comparing the promising output candidates with the oracle labels and updating the model weights accordingly. As a by-product of the avoidance of gradient computation, those methods have faster training speed compared with probabilistic gradient-based learning methods like CRF.

3. Method

In previous studies, a large scale of annotated data is used to train a loanword identification model. They treated the loanword detection as a sequence labeling problem. However, the annotated data for loanword identification are very difficult to obtain. So, one of the contributions of this study is the data augmentation for loanword identification. We propose to use a lexical constraint GAN to generate more sentences for loanword identification model training. Another contribution of this paper is the combination of several features for loanword identification model; we introduce three features such as embedding fusion feature (word level and character level), pronunciation similarity feature, and POS feature.

3.1. Overall Architecture

Our proposed method includes two parts:(1)Data augmentation for loanword identification.(2)Log-linear RNN-based loanword identification model.

To generate more training data for loanword identification, we propose a lexical constraint GAN-based data augmentation model. Recent methods on loanword identification often trained on features such as pronunciation similarity, POS similarity, and so on. However, these kinds of methods usually suffer from data sparseness or lack of semantic knowledge. To overcome this, we introduce a log-linear RNN-based loanword identification model which combines word-level and character-level embedding fusion features, pronunciation similarity, and POS features to predict Arabic, Chinese, Russian, and Turkish loanwords in Uyghur. The main idea of loanword identification in low-resource languages is as follows: we first use the data augmentation model to generate more training data for loanword identification in Uyghur; then, several features such as word- and character-level embedding features, pronunciation similarity, and POS features are proposed to build a multiple feature fusion-based loanword identification model (Figure 2).

3.2. Data Augmentation for Loanword Identification

Recent studies on loanword identification task often suffer from limitation of training data. In this study, we propose to use a lexical constraint GAN to generate more annotated data for the loanword identification task. As an extension of traditional GAN, our data augmentation model also includes two main parts: a generator and a discriminator. The difference is that we use two generators and a discriminator to build the data augmentation model for low-resource loanword identification. We introduce the details of our proposed model in this section.

3.2.1. Generators

We follow the work of [6] and extend the backward and forward generators to adapt to the loanword identification task. In our study, we use the loanwords of a specific language as the lexical constraint to generate more training data. Similar to [6], given a loanword, the backward generator takes it as the sentence’s starting point and generates the first half sentence backwards. Then, the sequence produced by the backward generator is reversed and fed into the forward generator. It then learns to generate the whole sentence with the aim of fooling the discriminator.

We can define the backward generator aswhere denotes a given loanword and indicates the length of generated training sentence. The generated sentence is . The backward generator generates the first half of the sentence, while another half of sentence is generated by the forward generator. and are parameters of the backward and forward generators.

The generator of the entire sentence can be defined aswhere and are descripted as above.

The two generators have the same structure but have distinct parameters. To improve the coherence of the constrained sentence, we employ an LSTM-based language model with dynamic attention mechanism (called attRNN-LM) as generator.

3.2.2. Discriminator

Another important component in our proposed method is the discriminator, which takes sentence pairs as input and distinguishes whether a given sentence pair is real or generated. It guides the joint training of two generators by assigning proper reward signals. This module can be a binary classifier or a ranker. Following previous methods [21], we use Text-CNN as the discriminator which outputs a probability indicating whether the input is generated by humans or machines in the experiment.

3.2.3. Data Augmentation Model

To train the data augmentation model effectively, we first pretrain the backward and forward generators by standard MLE loss. Different from [6], we sample a loanword in our loanword list as the lexical constraint rather than select it randomly. Then, we use two generators and the lexical constraint to generate the training sentence. The discriminator is trained based on real sentence as positive sample and sentences generated by generators as negative samples. The discriminator’s output is the probability that the generated sentence is written by humans. We use the discriminator’s output as the reward to encourage the two generators to work together to generate sentences which are indistinguishable from human-written sentence. To make the training stable and prevent the perplexity value skyrocketing, we apply teacher forcing to give the generators access to the gold-standard targets after each policy training step.

3.3. Multiple Feature Fusion-Based Loanword Identification

Loanword identification can be defined as a sequence labeling problem. However, different from a traditional sequence labeling task, loanword identification task can apply some additional knowledge such as semantic similarity, pronunciation similarity, and POS tagging. As the data augmentation can provide us more annotated data for model training, we propose to use a deep neural network model to identify loanword in low-resource settings. The principle feature we used is the fusion of word- and character-level features, which combines the word relation and pronunciation similarity in loanword identification. We also incorporate external features such as pronunciation similarity and POS information into our method. In this section, we first describe features used in our proposed method and then define the details of the loanword identification method.

3.3.1. Features

We use three kinds of features in our proposed method: the fusion feature, pronunciation similarity, and POS feature.

Fusion Feature. In loanword identification task, word co-occurrence often plays a very important role. For example, in the English sentence “Tiananmen square is the most famous tourist destination in Beijing,” the Chinese loanword “Tiananmen” is most related to the Chinese loanword “Beijing.” In previous work, word embedding can capture word similarity and word relations with other words in a sentence. Therefore, we apply self-attention to obtain word embedding in our study. The most important advantage of the self-attention is that it can model dependencies between words.

We use the dot-product attention in this study:where are query, key, and value vectors, respectively. It should be noted that the self-attention was obtained without scaling. We setand at time step , the word embedding at time based on self-attention can be defined as

The most important feature in loanword identification task is the pronunciation similarity between the word in receipt language and its corresponding word in donor language. As convolutional neural networks (CNNs) have been proven to capture the character-level information in NLP tasks, CNNs can process the sequences in the current receptive filed akin to the attention mechanism [22]. Meanwhile, we also use max pooling to capture character-level features. The way we use CNN in our proposed method can be defined as

We follow the study of [23] and use a CNN with a redundant position of input sequences masked to extract the character-level features. is the filter width set as 3. The convolution operation is denoted with , and the padded position of input sequences is set as 0.

Max means a max pooling operation. We use it to capture the significant features assigned with the highest value for a given filter. Therefore, in the time step , the character-level representation from local view is obtained as

To fuse the word-level and character-level features together, we propose to concatenate two features with automatic adjustment (Figure 3). The final fusion representation can be defined aswhere and are word-level and character-level features, respectively, and and are corresponding parameters.

Pronunciation Similarity Feature. Intuitively, we find that a loanword often has a similar pronunciation with its corresponding donor word. A sample method to detect loanwords is to use a string similarity algorithm to compute the string similarity scores between the candidate loanword and a list of words in donor language. Then, we rank the scores and take the word with the best score as the donor word. In loanword identification task, we first transform donor and receipt language texts into a same writing system. For example, in Chinese loanword identification in Uyghur, we first convert these two language texts into Latin. Then, we apply the most commonly used string similarity algorithm—minimum edit distance (MinED)—in our loanword identification task.where and are lengths of donor word list and receipt word list, respectively, and represent donor languages (Arabic, Chinese, Russian, and Turkish) and receipt language (in this study indicates Uyghur), is the loanword label of the th receipt word, and is the minimum edit distance of two words. To adapt the loanword identification task, we first conduct text normalization on all datasets, which transform a text into a canonical (standard) form. Then, we carry on morphological segmentation on morphologically rich languages, such as Uyghur, Russian, and Turkish.

POS Feature. As loanwords are often nouns, we propose a part-of-speech (POS) feature to further constrain the loanword identification model. We first pretrain POS tagging models for donor languages and receipt language. Considering both the language resource and performance, we select CRF as the framework of POS tagging model. As POS models are ready, if a word in receipt and its corresponding candidate donor word are all nouns, we set the POS features as 1.

3.3.2. Loanword Prediction Model

Log-linear models play a considerable role in statistics and machine learning. The most important reason we chose the log-linear model as the basic framework of our proposed loanword prediction model was because features can be easily added into it. Additionally, the log-linear model has been widely used in NLP tasks such as SMT and NMT.

To adapt the loanword prediction task and include rich features such as BiLSTM, POS, and semantic feature into the model, we use log-linear RNNs [24] as the basic framework in our task. Log-linear RNN is similar to a RNN model. It allows a more general form of input to the network at each time step; that is , instead of allowing only the latest symbol to be used as input, along with the condition , it now allows an arbitrary feature vector to be used as input; this feature vector is of fixed dimensionality and allows it to be computed in an arbitrary (but deterministic) way from the combination of the currently known prefix and the context . This is a relatively minor change, but one that usefully expands the expressive power of the network.

The hidden state at time in our loanword identification task can be defined as

We assume that we have a priori fixed a certain background function and also defined features defining a feature vector of fixed dimensionality .

Therefore, the loanword label of word can be defined as

During training of our proposed loanword identification model, we use the cross-entropy loss to optimize the performance of our model [25].

4. Experiments

In this section, we evaluate the effectiveness of our proposed method.

4.1. Data

To fully evaluate the effectiveness of our proposed model, we conduct Arabic, Chinese, Russian, and Turkish loanword identification in Uyghur. The datasets used in our experiments are listed in Table 1. We crawl these corpora from the Internet. Then, we annotate a small part with loanword label by hands. In all texts, we assure that each sentence includes at least one loanword.

Data typeSize

Sentences100, 780125, 085143, 290132, 500

To train the data augmentation model, we also collect some monolingual data from Internet for each language (Table 2).


Size (words)0.321.05 B1.70 B1.14 B1.49 B

4.2. Settings
4.2.1. Data Augmentation

We train the data augmentation model on datasets described in Table 2. We set the same hyperparameters for forward and backward generators. All generators include 2-layer char-level LSTMs with 1024 hidden units. The dimension of word embeddings is set to 1024; the batch size, dropout rate, threshold of element-wise gradient clipping, and initial learning rate of Adam optimizer are set to 128, 0.5, 5.0, and 0.001; layer normalization is also applied. We set both backward and forward generators to one layered word-level LSTM with 1024 hidden units when training on datasets described in Table 2. For the hyperparameters of the discriminator, the filter window size is set to be 3, 4, 5, 6, and 7, and the kernel number of each filter is 512. We set the batch size as 64 and the number of iterations as 5000.

4.2.2. Loanword Identification

We implemented the log-linear RNNs by ourselves. We also developed the extended version of edit distance algorithm to adapt the loanword identification task. For the POS feature, we first pretrained a Uyghur POS tagging model; then, we tagged all Uyghur sentences based on this model.

We compared our method with several strong baseline systems: Rule [1], CRF [2], BLSTM-CNN [3], and ClEmbedding [4].

4.3. Results on Data Augmentation

Results on data augmentation and size of training data can be found in Tables 3 and 4, respectively.







Size302, 480325, 790314, 208336, 852

4.4. Results on Loanword Identification

The results on loanword identification on different methods can be found in Table 5.

DonorModelLoanword identification results (%)
RF1F1 (+)

RussianRule (+)72.0472.8969.3170.1870.6571.28
CRF (+)71.6372.4567.2868.1569.3970.23
BLSTM-CNN (+)71.4572.2670.5071.3170.9771.78
ClEmbedding (+)73.1273.9471.8472.6272.4773.27
Ours (+)74.8075.6273.6474.2074.2274.90

ArabicRule (+)69.0569.8468.1769.0268.6169.43
CRF (+)69.8370.6567.4268.2968.6069.45
BLSTM-CNN (+)68.7069.5269.8570.6769.2770.09
ClEmbedding (+)72.9573.7672.0372.8572.4973.30
Ours (+)73.9174.6272.3573.0673.1273.83

TurkishRule (+)72.0272.8669.8770.5070.9371.66
CRF (+)71.4672.2969.0269.9570.2271.10
BLSTM-CNN (+)71.2572.0470.4371.1870.8471.61
ClEmbedding (+)72.9673.6473.0873.8573.0273.74
Ours (+)75.2476.0974.3675.1474.8075.61

ChineseRule (+)70.3271.1369.7770.5870.0470.85
CRF (+)70.8571.6469.2470.0570.0470.84
BLSTM-CNN (+)70.5871.3469.9870.7970.2871.06
ClEmbedding (+)71.6772.4871.3572.1471.5172.31
Ours (+)74.3075.0772.8873.9573.5874.51

5. Analysis

Table 3 presents experimental results on data augmentation for loanword identification. We can find that our proposed lexical constraint method achieves the best performance compared with other strong baseline systems in all evaluation metrics. The most important reason is that our method guarantees the fluency and semantic consistency of generated sentence at the same time. Table 4 shows the size of Uyghur sentence (with loanword in different donor languages) generated by our proposed data augmentation model. For loanwords in different donor languages, we obtain the largest Uyghur datasets with Turkish loanwords; one possible reason is that Uyghur and Turkish are closely related. We obtain the fewest sentences with Arabic; it is because Uyghur and Turkish have very different grammar and syntax.

The first part in Table 5 describes experimental results on different methods with the original training data. We found that the CRF and rule-based model outperform BLSTM-CNN method; one possible reason is the limitation of annotated data. Because the ClEmbedding model can exploit semantic information obtained from monolingual data, the ClEmbedding model achieves slightly better results compared with the CRF and rule-based model. Compared with other baseline models, our method incorporates word-level and character-level features pretrained from monolingual corpora into one model; therefore, our method achieves best results, but the improvement is not significant. This is because our method also suffers from data sparseness during model training.

The second part of Table 5(with (+)) presents loanword identification results on different methods with our generated training data (data augmentation). We can find that the generated training data improve all baseline models significantly. The CRF-based model has the ability of generalization, but the data sparseness still weakens the loanword identification performance significantly. The BLSTM-CNN + method also achieves better performance compared with the BLSTM-CNN. Both CRF+ and BLSTM-CNN + benefit from data augmentation. Although ClEmbedding + relies on monolingual data, it also obtains performance improvements due to loanword identification results are added. Our proposed method incorporates RNN features and external features into one model, so it achieves the best performance among all baseline systems.

Table 6 presents results on different features in our proposed method (we take Turkish and Chinese loanword identification as examples). We find that models with all features achieve best performance in both Turkish and Chinese loanword identification tasks. As for single feature, the fusion feature is more important than others; one possible reason is that the fusion feature combines word-level and character-level features at the same time. Except the fusion feature, pronunciation similarity feature outperforms other features because the pronunciation similarity is the most intuitive feature in loanword identification task. Although the POS cannot achieve comparative performance with others, we find that the combination features with POS always outperform others.

DonorFeature(s)Loanword identification results (%)
RR (+)F1F1 (+)

+fusion, pronun73.4074.2072.6473.4073.0273.80
+fusion, pos74.6375.4273.7074.5274.1674.97
+pronun, pos74.2575.0673.4574.2473.8574.65

+fusion, pronun72.4373.2571.0271.8471.7272.54
+fusion, pos73.6174.4072.2673.0272.9373.70
+pronun, pos73.2574.0371.9772.8972.6073.46

In Table 5, we describe results on different donor languages. We can easily find that our method achieves best performance on Turkish loanword identification task. One important reason is that Turkish and Uyghur belong to the same language family, and they share much vocabulary and grammar compared with other donor languages. Our model also achieves better results on Russian loanword identification than Chinese and Arabic; one possible reason is that Russian has a deep influence on Uyghur, and Uyghur is sometimes written in a Cyrillic alphabet, which is the basic writing system in Russian. Because people who can speak Uyghur can often speak Chinese fluently, Chinese has a significant impact on Uyghur. Although Uyghur and Arabic share the same writing system, two languages belong to different language families. So, Arabic loanword identification achieves the worst performance.

6. Conclusion

The main goal of this study is to improve the performance of loanword identification for low-resource language. Our contribution includes two parts: (1) data augmentation for loanword identification and (2) loanword identification based on multiple feature fusion. In particular, data augmentation alleviates the data sparseness occurring in the loanword identification model training; we optimize the loanword identification model by introducing several features such as fusion feature of word- and character-level embeddings, pronunciation similarity, and POS feature into one model based on a log-linear RNN. To evaluate the effectiveness of our proposed method, we conduct experiments on several baseline models. Experiments show that our proposed loanword identification method achieves the best performance.

In our future work, we plan to improve the robustness of the loanword identification model by generating more diverse training data and incorporating richer contextual information into it.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.


This research was funded by the National Natural Science Foundation of China (no. 61906158).


  1. R. T. McCoy and R. Frank, “Phonologically informed edit distance algorithms for word alignment with low-resource languages,” in Proceedings of the Society for Computation in Linguistics (SCiL’18), pp. 102–112, Salt Lake City, UT, USA, January 2018. View at: Google Scholar
  2. G. Akın Şeker and G. Eryiğit, “Extending a CRF-based named entity recognition model for Turkish well formed text and user generated content 1,” Semantic Web, vol. 8, no. 5, pp. 625–642, 2017. View at: Google Scholar
  3. C. Mi, Y. Yang, L. Wang, Xi Zhou, and T. Jiang, “A neural network based model for loanword identification in Uyghur,” in Proceedings of the 11th International Conference on Language Resources and Evaluation (LREC’18), Miyazaki, Japan, May 2018. View at: Google Scholar
  4. C. Mi, Y. Yang, L. Wang, Xi Zhou, and T. Jiang, “Toward better loanword identification in Uyghur using cross-lingual word embeddings,” in Proceedings of the 27th International Conference on Computational Linguistics, pp. 3027–3037, Santa Fe, NM, USA, August 2018. View at: Google Scholar
  5. C. Mi, L. Xie, and Y. Zhang, “Loanword identification in low-resource languages with minimal supervision,” ACM Transactions on Asian and Low-Resource Language Information Processing, vol. 19, no. 3, pp. 1–22, 2020. View at: Publisher Site | Google Scholar
  6. D. Liu, J. Fu, Q. Qu, and J. Lv, “BFGAN: backward and forward generative adversarial networks for lexically constrained sentence generation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 27, no. 12, pp. 2350–2361, 2019. View at: Publisher Site | Google Scholar
  7. Y. Tsvetkov and C. Dyer, “Lexicon stratification for translating out-of-vocabulary words,” in Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, vol. 2, Beijing, China, July 2015. View at: Google Scholar
  8. D. Gerz, I. Vulić, E. M. Ponti, R. Reichart, and K. Anna, “On the relation between linguistic typology and (limitations of) multilingual language modeling,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 316–327, Brussels, Belgium,, November 2018. View at: Google Scholar
  9. C. Mi, Y. Yang, X. Zhou, L. Wang, Li Xiao, and T. Jiang, “Recurrent neural network based loanwords identification in Uyghur,” in Proceedings of the 30th Pacific Asia Conference on Language, Information, and Computation: Oral Papers, pp. 209–217, Seoul, Korea, October 2016. View at: Google Scholar
  10. J. Liu, Y. Pan, F. X. Wu, and J. Wang, “Enhancing the feature representation of multi-modal MRI data by combining multi-view information for MCI classification,” Neurocomputing, vol. 400, pp. 322–332, 2020. View at: Google Scholar
  11. C. Amit, “A visual survey of data augmentation in NLP,” 2020, View at: Google Scholar
  12. X. Zhang, J. Zhao, and Y. LeCun, “Character-level convolutional networks for text classification,” 2015, View at: Google Scholar
  13. S. Edunov, M. Ott, M. Auli, and D. Grangier, “Understanding back-translation at scale,” 2018, View at: Google Scholar
  14. C. Coulombe, “Text data augmentation made simple by leveraging nlp cloud apis,” 2018, View at: Google Scholar
  15. H. Zhang, M. Cisse, Y. N. Dauphin, and L.-P. David, “Mixup: beyond empirical risk minimization,” 2017, View at: Google Scholar
  16. V. Kumar, A. Choudhary, and E. Cho, “Data augmentation using pre-trained transformer models,” 2020, View at: Google Scholar
  17. X. Sun, S. Ma, Y. Zhang, and X. Ren, “Towards easier and faster sequence labeling for natural language processing: a search-based probabilistic online learning framework (SAPO),” Information Sciences, vol. 478, pp. 303–317, 2019. View at: Publisher Site | Google Scholar
  18. F. Sha and F. Pereira, “Shallow parsing with conditional random fields,” in Proceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pp. 213–220, Edmonton, Canada, September 2003. View at: Google Scholar
  19. X. Sun, “Structure regularization for structured prediction,” NIPS, vol. 14, pp. 2402–2410, 2014. View at: Google Scholar
  20. S. V. N. Vishwanathan, N. N. Schraudolph, M. W. Schmidt et al., “Accelerated training of conditional random fields with stochastic gradient methods,” in Proceedings of the 23rd International conference on Machine learning, pp. 969–976, Pittsburgh, PA, USA, June 2006. View at: Google Scholar
  21. L. Yu, W. Zhang, J. Wang et al., “Seqgan: sequence generative adversarial nets with policy gradient,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, no. 1, 2017. View at: Google Scholar
  22. J. Liu, D. Zeng, R. Guo, M. Lu, F. X. Wu, and J. Wang, “Mmhge: detecting mild cognitive impairment based on multi-atlas multi-view hybrid graph convolutional networks and ensemble learning,” Cluster Computing, vol. 24, no. 1, pp. 103–113, 2021. View at: Google Scholar
  23. Z. Yang, H. Chen, J. Zhang et al., “Attention-based multi-level feature fusion for named entity recognition,” in Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}, pp. 3594–3600, Yokohama, Japan, January 2020. View at: Google Scholar
  24. M. Dymetman and C. Xiao, “Log-linear RNNs: towards recurrent neural networks with flexible prior knowledge,” 2016, View at: Google Scholar
  25. J. Cheng, J. Liu, H. Yue, H. Bai, Y. Pan, and J. Wang, “Prediction of glioma grade using intratumoral and peritumoral radiomic features from multiparametric MRI images,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, 2020. View at: Publisher Site | Google Scholar

Copyright © 2021 Chenggang Mi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.