Computational Intelligence and Neuroscience

Computational Intelligence and Neuroscience / 2020 / Article

Research Article | Open Access

Volume 2020 |Article ID 7834953 | https://doi.org/10.1155/2020/7834953

Nan Zhou, Junping Du, Zhe Xue, Chong Liu, Jinxuan Li, "Cross-Modal Search for Social Networks via Adversarial Learning", Computational Intelligence and Neuroscience, vol. 2020, Article ID 7834953, 12 pages, 2020. https://doi.org/10.1155/2020/7834953

Cross-Modal Search for Social Networks via Adversarial Learning

Academic Editor: Raşit Köker
Received20 Jan 2020
Revised01 May 2020
Accepted25 Jun 2020
Published11 Jul 2020

Abstract

Cross-modal search has become a research hotspot in the recent years. In contrast to traditional cross-modal search, social network cross-modal information search is restricted by data quality for arbitrary text and low-resolution visual features. In addition, the semantic sparseness of cross-modal data from social networks results in the text and visual modalities misleading each other. In this paper, we propose a cross-modal search method for social network data that capitalizes on adversarial learning (cross-modal search with adversarial learning: CMSAL). We adopt self-attention-based neural networks to generate modality-oriented representations for further intermodal correlation learning. A search module is implemented based on adversarial learning, through which the discriminator is designed to measure the distribution of generated features from intramodal and intramodal perspectives. Experiments on real-word datasets from Sina Weibo and Wikipedia, which have similar properties to social networks, show that the proposed method outperforms the state-of-the-art cross-modal search methods.

1. Introduction

With the rapid development of mobile networks and “we media” [1], cross-modal information search [2] has become a research hotspot. Users publish multimedia information on social network platforms such as Weibo and Twitter, where public opinion is expressed through natural language and visual information. Cross-modal information search meets users’ needs for data diversity, especially on social networks. Various types of topics (e.g., news, tips, and stories) occur in multimedia forms on social networks, conveying valuable information for various users, including common people, companies, and regulators. The most direct way to fulfill users’ diversified information needs is to maximally mine the resemblance and correlations of the information and present the content relevant to users’ queries [3, 4]. However, cross-modal correlation analysis faces the basic challenge of bridging the heterogeneity gap [5, 6] between different media, which is also a key issue for cross-modal search.

Bridging the heterogeneity gap in multimodal data, which feature different statistical characteristics, is the major issue in analyzing and processing multimodal datasets with intelligent technologies [7]. In general, some current research addresses the problem by constructing multiple nonlinear transformations [8] to build a common semantic subspace for multimodal data through deep learning [9]. With the subspace, the nonlinear transformations are learned to generate feature representations for correlation maximization [10]. The representative classical methods are canonical correlation analysis (CCA) [11] and variants such as deep CCA (DCCA) [12]. With the development of tabular learning and deep learning research, such strategic methods have gradually been divided into two groups: real-valued representations and binary-valued representations [13]. Other works focus on selecting relevant features that are, then, adopted to construct correlations from multimodal features to achieve cross-modal search through feature selection and matching [14, 15]. The methods used according to this strategy are designed to discover dense feature clusters with high similarity learned by algorithms for cross-modal data [16].

In addition, the semantic sparseness of cross-modal data from social networks results in misleading content in both the textual and visual modalities. Cross-modal data on social networks present characteristics that reflect many aspects of real-world events in quality-restricted forms [14]. The massive quantity of cross-modal data on social networks provides an opportunity to uncover relations between events and discover additional content related to the target event in a variety of media. The forms and characteristics of social network cross-modal data require many details of features such as local correlations to be mined and learned by intelligent algorithms. To overcome the semantic sparseness of cross-modal data from social networks, we adopt self-attention [17] to discover the differential importance of local semantic features according to the target topic throughout the global representation sensors. Self-attention can be used to assign weight values for different items in feature sequences to perceive significance. Li et al. [18] proposed a positional self-attention with contention (PSAC) architecture to capture long-range dependencies and position information. Through the application of self-attention to perceive significance, PSAC significantly outperforms its predecessor. Gao et al. [19] presented hierarchical LSTMs with an adaptive attention method to perceive the spatial-temporal attention for visual regions or frames to predict related words. This method with adaptive attention outperforms the previous state-of-the-art methods.

In this paper, we propose a cross-modal search method for social network data that capitalizes on adversarial learning. In addition, we adopt self-attention-based neural networks to generate modality-oriented representations for further inter-modal correlation learning. A search module is implemented based on adversarial learning, through which the discriminator is designed to measure the distribution of generated features from intramodal and intramodal perspectives. The discrimination is deployed as a compound neural network whose parameters are optimized under union losses following the adversarial learning mechanism to generate the most appropriate representations of cross-modal data features. The contributions of the paper are summarized as follows.(i)We propose a supervised cross-modal adversarial learning method integrated with self-attention. The method generates cross-modal representations following the original modality and topic label distributions from the perspective of social network data characteristics under the mechanism of self-attention.(ii)The proposed method incorporates local semantic features distributed as word groups in texts and blocks in images to maximize the cross-modal correlations based on adversarial learning.(iii)The part of the adversarial learning component in the designed adversarial learning framework is used effectively to rank the search results.

The unstandardized writing conventions of user-generated text and the frequently low quality of user-submitted images submitted on social networks result in semantic sparseness. Semantic sparseness is the main obstacle to cross-modal information search in social networks based on global semantic features. Our proposed method, cross-modal search with adversarial learning (CMSAL), integrates self-attention to explore local semantic features expressing key semantic features of the target topics. Words (in text) and pixel blocks (in images) conveying target topics are the local semantic features to be explored and mined. The generated representations integrated with the local semantic features constitute the semantic space for social network cross-modal information search. The designed maximum losses are optimized based on adversarial learning to promote the efficiency of the generated representations for cross-modal search. The learning method is trained iteratively with the representation-generating process from intramodal and intermodal perspectives. In classical generative adversarial networks (GANs) [20], the optimal discriminator is useless in most cases [21]. We reused the optimal intermodal and intramodal restriction to provide ranked search results based on distribution measures. In contrast to the existing methods, this paper takes the semantic sparseness of social network content into consideration for the specific task of cross-modal information search.

2.1. Social Network Cross-Modal Search

With the development of information and mobile networks, social network platforms are becoming the most important source for multimedia data [22]. Cross-modal search strategies on social networks can be classified into two main groups: common semantic subspace learning and feature selection and matching. For multimodal data from social networks conveying more information [23], intelligent technologies are needed to excavate latent correlations within massive and complex cross-modal datasets from social networks. Cai et al. [24] proposed a joint topic model to track and search target social information based on cross-modal feature sequence analysis and learning. Fang et al. [25] proposed a data transformation method to handle heterogeneous data for cross-modal event analysis and searches in social networks. Qing et al. [26] proposed an event and content search method based on automatic identification and tracking from a large amount of cross-modal data from social networks. Lee et al. [27] provided a common search framework for online social network hotspot events. The method normalizes the data content of different media based on the graph-based algorithm combination sorting event list for content normalization. It unifies the stream-based media data and the registration-based cross-media data, which realizes the cross-media search for the target event. Zhang et al. [28] studied the hierarchical information quad-tree index structure based on spatiotemporal characteristics, including temporal proximity, spatial proximity, and visual relevance. The method is also used to solve cross-modal search problems in social networks. Deng et al. [29] proposed a deep hash network based on triplets for cross-modal retrieval of social networks. The method uses a triple label to describe the relative relationship between the three instances as a supervisor to capture a more general semantic correlation between cross-modal instances.

Social network cross-modal search is related to the traditional cross-modal search on multimedia representation extraction and correlation analysis. Furthermore, cross-modal contents from social networks need to pay attention to global and local semantic associations in semantic sparseness, which is determined by the characteristics of the social network data. The emergence of GANs [20] provides a series of methods for semantic extractions and representations under sparse semantic conditions that are gradually applied to the field of cross-media search.

2.2. Adversarial Learning Cross-Modal Search

Recently, GANs [20] have been widely used because of their ability to learn and process visual and sequenced features. A series of approaches have been proposed to reduce the gap between different modalities based on adversarial learning of the statistical characteristics of the transformed features. Following this strategy, He et al. [30] introduced a cross-modal retrieval method based on unsupervised adversarial learning. The method constructed an adversarial learning feature transformation for the statistical properties on cross-modal search. Peng et al. [5] proposed a method for common cross-modal representation based on GAN. Through well-learned cross-modality representations, many applications such as cross-modal similarity matching can be conducted. Gu et al. [4] provided a GAN-based method incorporating corporate generative models into cross-modality embedding for cross-modal search. The method encouraged the textual features as the basis to generate an image similar to the ground truth, and vice versa for images to texts. Shang et al. [31] proposed a dictionary learning-based cross-modal search method. The method used a dictionary learned as a feature a reconstructor, cooperating with adversarial learning to mine cross-modality statistical characteristics. Wen et al. [32] proposed a cross-modal search method based on similarity transferring. The method uses adversarial learning to build a semantic structure in the common representation subspace for preserving the semantic structure between unpaired items across different modalities. Wang et al. [33] proposed an adversarial learning retrieval method that imposed triplet constraints for feature generation to minimize the heterogeneous gap of cross-modal data with the same semantic labels. The greatest advantage of adversarial learning is cross-modal synthesis. Gao et al. [34] presented a method named the perceptual pyramid adversarial network (PPAN) to synthesize photorealistic images and texts based on adversarial learning. The method is composed of a generator optimized with perceptual loss to obtain diverse images and a discriminator for multiple purposes, such as semantic consistency, image fidelity, and class invariance.

For other strategies, deep quantization and deep hashing based on adversarial learning are also used for cross-modal search. Yang et al. [35] proposed a method known as shared predictive deep quantization (SPDQ). In this method, a shared semantic subspace is defined for cross-modal features. The method builds a joint deep network architecture to exploit compact cross-modal representations. The method preserves intramodal and intermodal similarities in an efficient way. Deep hashing also follows the strategy to learn compact binary code for cross-modal similarity computation efficiency. Li et al. [36] presented a self-supervised adversarial hashing (SSAH) method. The method learns the high-dimensional features and hash codes for cross-modality information through two adversarial networks. The search similarity is maximized according to the semantic relevance in a highly computationally efficient manner.

In contrast to traditional methods of latent semantic subspace learning [37], cross-modal search based on GAN or adversarial learning takes advantage of the capacity for feature distribution construction and discrimination learning [33]. There are also many methods that adopt adversarial learning for hashing to realize cross-modal search [38, 39]. These methods convert the matching problem in cross-modal search to the Hamming distance calculation based on the multimedia effective binary representation. Such a calculation strategy improves the matching efficiency of cross-modal search. However, in the construction of binary representations, some semantic features of the original multimedia are lost. The proposed method in this paper focuses on local semantic feature extraction based on self-attention [17] and adversarial learning [20] to solve the problem of minimizing the heterogeneity gap for cross-modal data with the same semantic labels.

3. The Proposed CMSAL Method

3.1. Problem Definition

In general, we define cross-modal data as , 1 ≤ d ≤ D, meaning that there are D topics in the data domain on the amount. For each topic, related contents are expressed in the form of text and images as (1 ≤ m ≤ M, 1 ≤ n ≤ N). In each topic, there are M text instances and N image instances conveying the related semantic information to Cd labeled by ld. There are some special cases for (M ≥ 1, N = 0) and (M= 0, N ≥ 1), in which the problem degenerates into the unimodal case. Another case is (M = 1, N = 1). In this case, the situation agrees with most definitions in current works.

Raw text and images are preprocessed into representation features by word embedding [40] and VGGNet [41], according to the modality. The presentation features for texts and images are interfaces for further complex computing in the learning procedure. Let be the collection of cross-modal original features (word embedding features for text and CNN features for images) with the one-hot label vector yd for topic d, in which represents the word embedding feature for the m-th text entry under topic d.

For further correlation maximization learning, the presentation features are explored to extract local features that are sensitive to modality characteristics. The features convey the same semantics in word groups and image blocks represented as , meaning the k-th block word feature, which is the same as with K blocks as an empirical value. The extraction process is defined as . shortened as for text representation features with the parameters of , while shortened as ) for image features with the parameters of . and are the generation processes interacting with the discriminator to optimize parameters jointly by adversarial learning. A restriction is designed to measure the distribution of and from intramodal and intermodal aspects to guide the generation. and output more appropriate representation features by episodes. The general framework of the proposed method is illustrated in Figure 1.

3.2. Constructions of Cross-Modal Representation Feature Generation

Cross-modal representation feature generation is conducted to explore the local semantic relationships between features from different modalities and reconstruct the representations to reflect the relationships in computational matrixes. The procedure is designed under a supervised representation learning mechanism in which self-attention is adopted. Taking text modality as an example, , , and are the functions to transform the original features (word features for text in fixed-size blocks) into a subspace as follows:where means the k-th text block word embedding feature of a text document on topic d. , , and are the parameter vectors of , , and . Similarly, , , and are the corresponding functions for the image modality with the parameter vectors , , and .

The original features of the two modalities are cut into fixed-size blocks. In general, we cut the original feature into K blocks. The blocks of original text features are composed of word vectors, while the blocks of original image features cover the CNN features of pixels. For example, the attention between the i-th and the j-th blocks is calculated as follows:where indicates the model attention parameter related to the j-th feature block when generating the representation features of the i-th block in the specific word embedding feature of the corresponding text on topic d. Similar to image modality, is used for images in CNN feature blocks. For the i-th block of a specific text piece of content, the representation features can be presented as follows:

The representation features of a whole text about the topic d can be presented as , which is the same as . It is regarded as a global semantic representation. The value of K is a hyperparameter determined by experiences and data contexts. In the experiment, we set the value of K according to the corresponding original cross-modality features. Otherwise, the value of K also determines the sizes of (), (), and () as parameters. However, it will have little impact on the actual representations through cross-modal presentative feature generation.

3.3. Learning Metric for the Proposed Method

In this section, we propose the generation and discrimination losses to train the proposed CMSAL. The generation loss guides the representation features generation and consists of a label loss and a similarity loss. The label loss aims to minimize the distribution difference between the representation features and corresponding topic semantic labels. The similarity loss is used to minimize the distance among the intermodal samples about the same topic. These two loss terms are defined as the generation loss for guiding the representation features generating procedure. The discrimination loss is defined to distinguish modalities. The multiple losses are collaborated into a minimax loss to optimize the generation of representation features for appropriate cross-modal search features.

3.3.1. The Generation Loss

The generation loss is decomposed into two loss terms: the label loss and the similarity loss. The label loss ensures that the distributions generated representation features following those of semantic topics. The loss is presented aswhere and are the topic labels for corresponding features in the form of a one-hot vector. The symbol is the function to predict topic probability distribution for each text or image term of the representation features. M and N are the amounts of the original features for text and images, respectively. As described in Section 3.1, we conduct the collection based on M = N for a clear expression and thinking. Therefore, equation (4) can be further expressed as follows:

The label loss guides the training of the parameters of θt and to generate representation features following the topic distribution of corresponding samples. The label loss is the intramodal loss used to maintain the intramodal data correlations. Based on the premise of M = N, the similarity loss is defined as follows:

The similarity loss acts as the intermodal loss to maximize correlations between cross-modal samples with the same topic distribution by closing the distance difference of representation features and topic labels.

The losses presented in equations (5) and (6) are the basics to guide representation feature generation by supervised learning for adjusting the parameters of the networks. As parts of the generation loss, the label loss and the similarity loss are integrated by weighted summation presented as equation (9).where α and β represent the contribution weights of the corresponding deviation values to the loss function, through which the optimization of generation loss is directly affected by the two empirical values.

3.3.2. The Discrimination Loss

The discrimination of the method is the key component to realize cross-modal adversarial learning. It aims to discriminate the modalities for the constructions of representation features about the same topic. We define the discrimination loss as follows:where mi is the modality label as a one-hot vector and aims to map the generated representation features into the modality discrimination space under the parameter θp. Different from the generation loss, the discrimination loss promotes representation feature generation indirectly. The generator will output more appropriate representation features by parameter optimization and adversarial learning with a discriminator.

3.4. The Adversarial Training Procedure

To ensure the correlation maximum of cross-modal representation features for the same topic distribution, cross-modal representation feature generation and intermodal discrimination interact with adversarial learning. We construct the minimax game [20] as follows:where , , and are optimized values for the joint losses. The minimax game will minimize generation loss and maximize the discrimination loss. The generation loss is going to construct cross-modal representation features to maximize relationships for the same semantic topic distribution. The discrimination loss will distinguish modality discrepancies. The parameters θp are fixed for optimizing θt and during the minimization procedure, while θt and are fixed for optimizing θp during the maximization procedure.

3.5. The Matching Similarity for CMSAL

Based on the optimized parameters, the cross-modal representation features constitute the correlation maximized representation space for text word embedding features and image CNN features. The generated representation features from cross-modal features are used to calculate similarities for cross-modal matching to search. The L2-norm is used to calculate the similarity presented as follows:

As presented in equation (6), the similarity calculation is included in the similarity loss. The similarity calculation is based on the optimized parameters and for appropriate results. The matching algorithm is shown in Algorithm 1.

Input: Query set Q = {q1, q2, …, qT} about the target topic d; cross-modal presentative data features from social networks generated by the networks with optimized parameters and :
Output: Search result list.
(1)For q in the query set Q:
(2) Distinguish modality type of q
(3) Preprocessing q into corresponding features blocks as for q in text or for q in images.
(4)   Extracting representation features: based on the optimized parameters for q in text or based on the optimized parameters .
(5)  For s in the cross-modal representation features set Sdoc:
(6)   Computing the similarity according to the query similarity = sim (Sq, s)
(7) End For
(8)End For

Sorting and picking up the top K similarities are executed as the evaluation scope with the corresponding representation features. The corresponding content of the representation features in a list is returned according to the sorted top K similarities as the evaluation scope. The algorithm outputs cross-modal search results according to the query. The matching similarities are calculated based on the trained proposed method to obtain the most appropriate results.

4. Experiments and Analyses

Experiments on real-world datasets are conducted to verify the effectiveness of the proposed method on cross-modal search from social networks. The real-world datasets consisted of text-image pairs collected from Sina Weibo. Without loss of generality, the widely used Wikipedia [42] and NUS-WIDE [43] cross-modal datasets are also used to verify the effectiveness of the proposed method. In this section, the effects of changing empirical hyperparameter values and cross-modal search efficiency are shown and analyzed.

4.1. Experimental Setup
4.1.1. Dataset Descriptions

The data collected from Sina Weibo of 4735 text-image pairs are about four security event topics from June 10, 2012, to September 7, 2016. There are 2866 text-image pairs of 10 categories in the Wikipedia dataset. In the NUS-WIDE dataset, there are 9000 text-image pairs in 350 categories. Following the traditional machine learning strategy, 70% of the data are used as a training set, while the rest are used as a test set for the two datasets.

4.1.2. Evaluation Metrics

The mean average precision (MAP) for the top K and precision-scope curve are adopted as evaluation metrics to measure the performance of the proposed method. Following [33], MAP can be calculated as follows:

In equation (11), Q is the number of queries. K is the amount of the contents to be searched for results. The top k search precision is denoted as , which is also adopted as a measure for the search results for the scope K presented as a precision-scope curve. The average precision is computed in equation (11) as a component of equation (12).

4.1.3. Baselines

We compare the proposed CMSAL method with state-of-the-art methods on the Sina Weibo dataset and Wikipedia dataset. The selected methods are representative from classical applications to adversarial learning, such as JFSSL [44], CMDN [45], DCCA [12], ACMR [33], and CM-GAN [5].

4.1.4. Parameter Learning Results and Analyses

We conduct an experiment to show the impact of the empirical values α and β in equation (6) for the searching performance, of which the results will provide a basis for setting the empirical values in return. MAP is used to evaluate the performance while the empirical values vary. The evaluations of the two datasets are presented in Figures 24. The empirical values of alpha and beta are the corresponding weight parameters for the label loss and the similarity loss.

As shown in Figure 2, we evaluate the top 50 search results based on computing MAP@50 for varying alpha and beta on the Sina Weibo dataset. The MAP@50 value shows different distributions with the common point that MAP@50 obtains a better situation when beta = 0.1. This means that the similarity loss requires a smaller weight value than alpha for a high MAP@50 evaluation. As shown in Figure 3, the effects of empirical values for searching performance on the Wikipedia dataset are smaller than those of the Sina Weibo dataset. Different from Figure 2, there is less fluctuation of MAP@50 varying the values of alpha and beta. The results presented in Figure 3 also provide a reference for the alpha and beta. Considering the situations of Figures 2 and 3, empirical values can be set with a group of suitable values for appropriate search results.

Figure 4 presents the empirical values impacting the cross-modal search based on the NUS-WIDE dataset. The numerical distribution is relatively flat, as in Figure 3 for the NUS-WIDE dataset. The results show that the dataset property has a direct impact on the empirical value assignment. Similar to the Wikipedia dataset, the semantics of the cross-modal information NUS-WIDE dataset are more obvious with less sparsity. Furthermore, the correspondence of cross-modal data in NUS-WIDE is clearer by using simple text content as a semantic label. Therefore, the empirical values impacting the image-to-text search performance of MAP@50 in the NUS-WIDE dataset are greater than those in the Wikipedia dataset.

The proposed method sets the empirical values of alpha and beta according to the dynamic evaluations as described. The learning process is inseparable from appropriate empirical values. We incorporated appropriate values of alpha = 1 and beta= 0.1 for image searches with text input and alpha = 0.1 and beta= 0.1 text searches with image input in both the Sina Weibo dataset and the Wikipedia dataset. According to Figure 4, alpha = 0.1 and beta= 100 for image searches with text input and alpha = 100 and beta= 10 for text searches with image input will be appropriate for the NUS-WIDE dataset.

4.2. Search Result Evaluations and Analyses
4.2.1. MAP Evaluations and Analyses

Based on the appropriate values of alpha and beta for the learning process of the proposed CMSAL method, evaluations for search results are presented. In this section, we use MAP to show searching performances for the top 5, top 20, and top 50 results of CMSAL compared with the baseline methods. The evaluations on the Sina Weibo dataset are shown in Table 1, while those on the Wikipedia dataset are presented in Table 2. The evaluations on the NUS-WIDE dataset are presented in Table 3.


Traditional-basedDNN-basedGAN-based
JFSSLCMDNDCCAACMRCMGANCMSAL

MAP@5txt2img0.64780.71830.38850.86530.87770.8898
img2txt0.53510.58140.32510.71330.72570.9481
average0.59150.64990.35680.78930.80170.919

MAP@20txt2img0.61280.67990.35830.82380.83620.8539
img2txt0.51810.58430.32390.70710.71950.9412
average0.56550.63210.34110.76550.77790.8975

MAP@50txt2img0.51970.59060.32130.70650.71890.8353
img2txt0.52820.57290.31990.69920.71160.9293
average0.52390.58170.32060.70290.71530.8823


Traditional-basedDNN-basedGAN-based
JFSSLCMDNDCCAACMRCMGANCMSAL

MAP@5txt2img0.26850.44060.50940.62250.66290.6563
img2txt0.21510.34730.41250.49870.53910.5123
average0.24180.39400.46090.56060.60100.5843

MAP@20txt2img0.28310.42640.48950.61090.65130.6463
img2txt0.22090.35760.41020.49860.53900.5095
average0.2520.3920.44980.55480.59510.5779

MAP@50txt2img0.25430.41460.46240.57320.61360.6315
img2txt0.21780.34540.39560.48350.52390.5031
average0.23610.38000.42900.52840.56870.5673


Traditionally-basedDNN-basedGAN-based
JFSSLCMDNDCCAACMRCMGANCMSAL

MAP@5txt2img0.24640.41870.51100.63970.68720.8023
img2txt0.21970.33130.46820.48380.47970.6329
average0.13370.39080.58770.68840.62480.7474

MAP@20txt2img0.25110.54690.47920.60770.62450.6796
img2txt0.22310.49650.50500.52840.50420.5887
average0.24500.36990.40120.62490.53140.6640

MAP@50txt2img0.30150.40110.56700.65000.63010.6798
img2txt0.10010.47000.39100.48010.54660.4836
average0.25870.46970.50350.66210.66530.7976

In Table 1, txt2img means entering a text query with the target topic to search from images with the same topics (img2txt means the reverse). As shown, the proposed CMSAL method outperforms the selected baseline methods. For CMSAL itself, the task of img2txt obtains better evaluations on MAP for the top 5 than those of the txt2img task. The reason for this situation is that original images contain abundant semantic information that will be extracted and represented appropriately. The extracted CNN features can preserve and present the valuable local semantics in detail completely, while semantic units of text are simpler for the same purpose. It will be more reliable when querying an image, with more semantic information for searching text with target semantics.

Table 2 shows MAP evaluations on the Wikipedia dataset. Compared with Table 1, the evaluation values on the Wikipedia dataset show lower values than those on the Sina Weibo dataset (in Table 1). The Sina Weibo dataset contains typical raw real-world data from various users, including casual written text and low-resolution images, which provide sparse cross-modal semantics. As expected, the results on the Sina Weibo dataset achieved higher evaluation values than the results on the Wikipedia dataset. The reason for this situation is that semantic features in the Sina Weibo dataset are relatively concentrated and prominent.

As shown in Table 3, the proposed CMSAL method outperforms the selected standard methods. MAP evaluations on the NUS-WIDE dataset are smaller than those on the Sina Weibo dataset. The main reason is that the characteristics of the NUS-WIDE dataset are different from those of the Sina Weibo and Wikipedia datasets. On the NUS-WIDE dataset, images are labeled with relatively simple text content, which clarifies the correspondence between text and images. In addition, in terms of image data quality, the NUS-WIDE dataset has simplified semantic information as public datasets. Therefore, the MAP evaluations of search results on the NUS-WIDE dataset are closer to those on Wikipedia datasets.

4.2.2. Precision-Scope Evaluations and Analyses

As precision-scope curves are an indispensable form of evaluation for information search experiments, precision-scope curves of the proposed method CMSAL, and all the selected baseline methods. The experimental results on the Sina Weibo, Wikipedia, and NUS-WIDE datasets are shown in Figures 57.

As shown in Figures 5 and 6, the proposed CMSAL method shows a better performance than any of the other methods. In general, the measures of all the methods show similar trends with a small numerical gap. Similar to MAP evaluations, GAN-based methods achieve better performances than deep neural network- (DNN-) based methods which rely on targeted adversarial learning integrated with the advantages of DNNs. The classical DCCA method shows the worst values of the evaluation working in concert with MAP evaluations. The processing of nonlinear mapping and canonical correlation analysis learning is relatively independent for DCCA. However, the GAN-based method overcomes the disadvantages of traditional and DNN-based methods. The proposed method conducts appropriate representation feature generation to maximize correlations in adversarial learning. The results of the precision-scope curves demonstrate the effectiveness of the proposed method.

Figure 7 presents evaluations of precision-scope curves on the NUS-WIDE dataset for the tasks of searching for images from text input and searching for text from image input. As presented in Figure 7, the proposed CMSAL method outperforms other selected baseline methods. In addition, the precision-scope curves of CMSAL on the NUS-WIDE dataset outperform those on the Sina Weibo and Wikipedia datasets. The reason is that the cross-modal content in the NUS-WIDE dataset is simple and clear. As the semantic labels of images, the text has clear semantic features; thus, the tasks of text-to-image and image-to-text search show good computing properties in local semantic mining and matching for CMSAL.

5. Conclusions

In this paper, we propose a cross-modal search method for social network cross-modal data based on adversarial learning (CMSAL). The proposed method integrates self-attention based on adversarial learning to realize the cross-modal search for the social network. The method explores cross-modal semantic features from the perspective of global representations of images and texts for a specific topic. Through adversarial learning, the method reconstructs representations for cross-modal matching. The designed adversarial learning framework is effectively used to rank the search results. Experimental results validate the effectiveness of the proposed method.

Data Availability

The Sina Weibo dataset used to support the findings of this manuscript is collected with the help of Sina Weibo application interface from the website https://open.weibo.com/wiki/%E5%BE%AE%E5%8D%9AAPI?sudaref=www.baidu.com&display=0&retcode=6102. The Wikipedia dataset used to support the findings of this manuscript is taken from the website http://www.svcl.ucsd.edu/projects/crossmodal/. The NUS-WIDE dataset is taken from the website https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/NUS-WIDE.html.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported in part by the National Key R&D Program of China (no. 2018YFB1402600) and in part by the National Science Foundation of China (NSFC) under Grant nos. 61532006, 61772083, 61877006, and 61802028.

References

  1. Y. Wu, Z. Tong, C. Liu et al., “Research on the impact of we media marketing on brand communication,” in Proceedings of the 2018 2nd International Conference on Management, Education and Social Science (ICMESS 2018), Atlantis Press, Qingdao, China, June 2018. View at: Google Scholar
  2. F. Kou, J. Du, W. Cui et al., “Common semantic representation method based on object attention and adversarial learning for cross-modal data in IoV,” IEEE Transactions on Vehicular Technology, vol. 68, no. 12, pp. 11588–11598, 2019. View at: Publisher Site | Google Scholar
  3. L. Zhang, B. Ma, G. Li, Q. Huang, and Q. Tian, “Generalized semi-supervised and structured subspace learning for cross-modal retrieval,” IEEE Transactions on Multimedia, vol. 20, no. 1, pp. 128–141, 2018. View at: Publisher Site | Google Scholar
  4. J. Huang, J. Cai, S. R. Joty et al., “Look, imagine and match: improving textual-visual cross-modal retrieval with generative models,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7181–7189, Salt Lake City, UT, USA, 2018. View at: Google Scholar
  5. Y. Peng and J. Qi, “CM-GANS: Cross-modal generative adversarial networks for common representation learning,” ACM Transactions on Multimedia Computing, Communications, and Applications, vol. 15, no. 1, pp. 1–24, 2019. View at: Publisher Site | Google Scholar
  6. Y. Peng, J. Qi, X. Huang, and Y. Yuan, “CCL: Cross-modal correlation learning with multigrained fusion by hierarchical network,” IEEE Transactions on Multimedia, vol. 20, no. 2, pp. 405–420, 2018. View at: Publisher Site | Google Scholar
  7. X. Yuan, L. He, H. Lu, L. Gao, and Y. Ji, “Deep adversarial metric learning for cross-modal retrieval,” World Wide Web, vol. 22, no. 2, pp. 657–672, 2019. View at: Publisher Site | Google Scholar
  8. P. Gao, D. Peng, X. Wang, and Y. Xiang, “Multimodal adversarial network for cross-modal retrieval,” Knowledge-Based Systems, vol. 180, pp. 38–50, 2019. View at: Publisher Site | Google Scholar
  9. P. Xiang, Z. Lin, Z. Yang et al., “Deep semantic space with intra-class low-rank constraint for cross-modal retrieval,” in Proceedings of the 2019 on International Conference on Multimedia Retrieval, pp. 226–234, Ottawa, ON, Canada, June 2019. View at: Google Scholar
  10. J. Yu and X. J. Wu, “Cross-modal subspace learning with Kernel correlation maximization and Discriminative structure preserving,” 2019, http://arxiv.org/abs/1904.00776. View at: Google Scholar
  11. M. Greenacre, Correspondence Analysis in Practice, Chapman and Hall/CRC, Boca Raton, FL, USA, 2017.
  12. G. Andrew, R. Arora, J. Bilmes et al., “Deep canonical correlation analysis,” in Proceedings of the International Conference on Machine Learning, pp. 1247–1255, Atlanta, GA, USA, 2013. View at: Google Scholar
  13. R. Liu, Y. Zhao, L. Zheng et al., “A new evaluation protocol and benchmarking results for extendable cross-media retrieval,” 2017, http://arxiv.org/abs/1703.03567. View at: Google Scholar
  14. L. Chu, Y. Zhang, G. Li, S. Wang, W. Zhang, and Q. Huang, “Effective multimodality fusion framework for cross-media topic detection,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 26, no. 3, pp. 556–569, 2016. View at: Publisher Site | Google Scholar
  15. X. Wang, A. Shimada, R. Taniguchi et al., “Coupled dictionary learning and feature mapping for cross-modal retrieval,” in Proceedings of the 2015 IEEE International Conference on Multi-Media and Expo (ICME), pp. 1–6, IEEE, Turin, Italy, 2015. View at: Google Scholar
  16. C. Kang, S. Xiang, S. Liao, C. Xu, and C. Pan, “Learning consistent feature representation for cross-modal multimedia retrieval,” IEEE Transactions on Multimedia, vol. 17, no. 3, pp. 370–381, 2015. View at: Publisher Site | Google Scholar
  17. H. Zhang, I. Goodfellow, D. Metaxas et al., “Self-attention generative adversarial networks,” 2018, http://arxiv.org/abs/1805.08318. View at: Google Scholar
  18. X. Li, J. Song, L. Gao et al., “Beyond RNNs: positional self-attention with co-attention for video question answering,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 1, pp. 8658–8665, 2019. View at: Publisher Site | Google Scholar
  19. L. Liu, X. Li, J. Song, and H. T. Shen, “Hierarchical LSTMs with adaptive attention for visual captioning,” IEEE Transactions on Pattern Analysis and Machine Intelligence, p. 1, 2019. View at: Publisher Site | Google Scholar
  20. I. Shen, J. Pouget-Abadie, M. Mirza et al., “Generative adversarial nets,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 2672–2680, 2014. View at: Google Scholar
  21. J. Su, “O-GAN: extremely concise approach for auto-encoding generative adversarial networks,” 2019, http://arxiv.org/abs/1903.01931. View at: Google Scholar
  22. T. Liu, F. Xue, J. Sun et al., “A survey of event analysis and mining from social multimedia,” Multimedia Tools and Applications, pp. 1–18, 2019. View at: Publisher Site | Google Scholar
  23. M. Liu, W. Guan, J. Yan, and H. Hu, “Correlation identification in multimodal weibo via back propagation neural network with genetic algorithm,” Journal of Visual Communication and Image Representation, vol. 60, pp. 312–318, 2019. View at: Publisher Site | Google Scholar
  24. H. Hu, Y. Yang, X. Li et al., “What are popular: exploring twitter features for event detection, tracking and visualization,” in Proceedings of the 23rd ACM International Conference on Multimedia, pp. 89–98, Brisbane, Australia, October 2015. View at: Google Scholar
  25. M. Fang, Y. Li, Y. Hu, S. Mao, and P. Shi, “A unified semantic model for cross-media events analysis in online social networks,” IEEE Access, vol. 7, pp. 32166–32182, 2019. View at: Publisher Site | Google Scholar
  26. S. Mao, T. Zhang, and C. Xu, “Online multimodal multiexpert learning for social event tracking,” IEEE Transactions on Multimedia, vol. 20, no. 10, pp. 2733–2748, 2018. View at: Publisher Site | Google Scholar
  27. W. Y. Lee, W. H. Hsu, and S. Satoh, “Learning from cross-domain media streams for event-of-interest discovery,” IEEE Transactions on Multimedia, vol. 20, no. 1, pp. 142–154, 2018. View at: Publisher Site | Google Scholar
  28. C. Zhang, R. Chen, L. Zhu, A. Liu, Y. Lin, and F. Huang, “Hierarchical information quadtree: efficient spatial temporal image search for multimedia stream,” Multimedia Tools and Applications, vol. 78, no. 21, pp. 30561–30583, 2019. View at: Publisher Site | Google Scholar
  29. C. Deng, Z. Chen, X. Liu, X. Gao, and D. Tao, “Triplet-based deep hashing network for cross-modal retrieval,” IEEE Transactions on Image Processing, vol. 27, no. 8, pp. 3893–3903, 2018. View at: Publisher Site | Google Scholar
  30. L. Gao, X. Xu, H. Lu et al., “Unsupervised cross-modal retrieval through adversarial learning,” in Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), pp. 1153–1158, IEEE, Hong Chong, China, July 2017. View at: Google Scholar
  31. F. Shang, H. Zhang, L. Zhu, and J. Sun, “Adversarial cross-modal retrieval based on dictionary learning,” Neurocomputing, vol. 355, pp. 93–104, 2019. View at: Publisher Site | Google Scholar
  32. X. Sun, Z. Han, X. Yin et al., “Adversarial cross-modal retrieval via learning and transferring single-modal similarities,” 2019, http://arxiv.org/abs/1904.08042. View at: Google Scholar
  33. B. Wang, Y. Yang, X. Xu et al., “Adversarial cross-modal retrieval,” in Proceedings of the 25th ACM international conference on Multimedia, pp. 154–162, Mountain View, CA, USA, 2017. View at: Google Scholar
  34. L. Gao, D. Chen, J. Song, X. Xu, D. Zhang, and H. T. Shen, “Perceptual pyramid adversarial networks for text-to-image synthesis,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, pp. 8312–8319, 2019. View at: Publisher Site | Google Scholar
  35. E. Xu, C. Deng, C. Li, W. Liu, J. Li, and D. Tao, “Shared predictive cross-modal deep quantization,” IEEE Transactions on Neural Networks and Learning Systems, vol. 29, no. 11, pp. 5292–5303, 2018. View at: Publisher Site | Google Scholar
  36. C. Liu, C. Deng, N. Li et al., “Self-supervised adversarial hashing networks for cross-modal retrieval,” in Proceedings of the Computer Vision and Pattern Recognition, pp. 4242–4251, Salt Lake City, UT, USA, 2018. View at: Google Scholar
  37. J. Wu, Z. Lin, and H. Zha, “Joint latent subspace learning and regression for cross-modal retrieval,” in Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 917–920, Tokyo, Japan, August 2017. View at: Google Scholar
  38. J. Song, T. He, L. Gao et al., “Binary generative adversarial networks for image retrieval,” in Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, February 2018. View at: Google Scholar
  39. X. Xu, F. Shen, Y. Yang, H. T. Shen, and X. Li, “Learning discriminative binary codes for large-scale cross-modal retrieval,” IEEE Transactions on Image Processing, vol. 26, no. 5, pp. 2494–2507, 2017. View at: Publisher Site | Google Scholar
  40. O. Shen and Y. Goldberg, “Neural word embedding as implicit matrix factorization,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 2177–2185, Montreal, Canada, 2014. View at: Google Scholar
  41. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” 2014, http://arxiv.org/abs/1409.1556. View at: Google Scholar
  42. N. Rasiwasia, J. Costa Pereira, E. Coviello et al., “A new approach to cross-modal multimedia retrieval,” in Proceedings of the 18th ACM International Conference on Multimedia, pp. 251–260, Firenze, Italy, 2010. View at: Google Scholar
  43. T. Chua, J. Tang, R. Hong et al., “NUS-WIDE: a real-world web image database from National University of Singapore,” in Proceedings of the Conference on Image and Video Retrieval, Santorini Island, Greece, July 2009. View at: Google Scholar
  44. K. Wang, R. He, L. Wang et al., “Joint feature selection and subspace learning for cross-modal retrieval,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 10, pp. 2010–2023, 2016. View at: Publisher Site | Google Scholar
  45. Y. Peng, X. Huang, and J. Qi, “Cross-media shared representation by hierarchical learning with multiple deep networks,” in Proceedings of the 2016 International Joint Conference on Artificial Intelligence, pp. 3846–3853, New York, NY, USA, July 2016. View at: Google Scholar

Copyright © 2020 Nan Zhou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views101
Downloads122
Citations

Related articles

We are committed to sharing findings related to COVID-19 as quickly as possible. We will be providing unlimited waivers of publication charges for accepted research articles as well as case reports and case series related to COVID-19. Review articles are excluded from this waiver policy. Sign up here as a reviewer to help fast-track new submissions.