skip to main content
research-article
Open Access

Self-supervised Short-text Modeling through Auxiliary Context Generation

Published:12 April 2022Publication History

Skip Abstract Section

Abstract

Short text is ambiguous and often relies predominantly on the domain and context at hand in order to attain semantic relevance. Existing classification models perform poorly on short text due to data sparsity and inadequate context. Auxiliary context, which can often provide sufficient background regarding the domain, is typically available in several application scenarios. While some of the existing works aim to leverage real-world knowledge to enhance short-text representations, they fail to place appropriate emphasis on the auxiliary context. Such models do not harness the full potential of the available context in auxiliary sources. To address this challenge, we reformulate short-text classification as a dual channel self-supervised learning problem (that leverages auxiliary context) with a generation network and a corresponding prediction model. We propose a self-supervised framework, Pseudo-Auxiliary Context generation network for Short-text Modeling (PACS), to comprehensively leverage auxiliary context and it is jointly learned with a prediction network in an end-to-end manner. Our PACS model consists of two sub-networks: a Context Generation Network (CGN) that models the auxiliary context’s distribution and a Prediction Network (PN) to map the short-text features and auxiliary context distribution to the final class label. Our experimental results on diverse datasets demonstrate that PACS outperforms formidable state-of-the-art baselines. We also demonstrate the performance of our model on cold-start scenarios (where contextual information is non-existent) during prediction. Furthermore, we perform interpretability and ablation studies to analyze various representational features captured by our model and the individual contribution of its modules to the overall performance of PACS, respectively.

Skip 1INTRODUCTION Section

1 INTRODUCTION

Short-text classification is a useful but challenging problem in various application settings such as sentiment analysis [40], dialogue systems [19], short-text topic modeling [35], and user intent detection [13]. Unlike paragraphs or documents, short text is ambiguous primarily due to the lack of context. The short text derives its context from real-world knowledge bases. A few examples of such cases are given in Table 1. The entity Tim Boyle relates to sports given the context “NFL” and “offseason.” Humans leverage existing knowledge bases to enhance their comprehension of short text.1 Moreover, they do not focus on the entire knowledge base but rather focus their attention toward a specific set of entities and their relationships to retrieve the relevant context. Current approaches in short-text classification lack such focus mechanisms and employ the entire knowledge base toward downstream tasks. This technique is useful when the downstream task depends on the entire knowledge base; e.g., email communications have limited explicit context of their topics. However, in several real-world applications, we notice adequate availability of auxiliary information that explicitly connects to the short text.

Table 1.
Short TextAuxiliary ContextLabel
Tim Boyle ready to take another step\( \ldots \) the offseason in \( \ldots \) the NFL \( \ldots \) a nose-down offseason\( \ldots \)Sports (News)
The Lord of the Rings: The Return of the King\( \ldots \)forces of good and evil fighting\( \ldots \) their quest to \( \ldots \)Adventure (Movies)
Belkin WaveRest Gel Mouse Pad\( \ldots \)comfortable with \( \ldots \)smooth, durable surface \( \ldots \)Office (Reviews)
  • Highlighted words significantly contribute to the final classification.

Table 1. Example of Short Text and Its Corresponding Auxiliary Context with the Class Label

  • Highlighted words significantly contribute to the final classification.

We discuss some applications in Table 1 where short text appears along with corresponding auxiliary information. For instance, the goal in the news classification task is to predict the category given the news headline. The description of the news and their authors serve as the auxiliary information. In the case of movie classification, predicting the genre from the movie title is the objective. The auxiliary information here is the synopsis, cast, and reviews. For the reviews on e-commerce platforms [6], the goal is to predict the product category from the review information. The auxiliary information here is the product type, description, and reviews based on historical purchases of the popular queries. In this article, we address this challenge of leveraging the auxiliary context information and enriching the representations of the short-text classification model.

Popular text classification approaches rely on semantic relations between words and their corresponding context. For training a language model, BERT [8] and XLNet [44] adopt a masking and a word-order permutation task, respectively. Although these models are effective in modeling sequences, there is a disconnect between the model training and the prediction phases. In other words, if BERT and XLNet are pre-trained on next sentence prediction, which contains a [SEP] tag to differentiate sentences, the fine-tuning would rely on the [SEP] tag in the data to distinguish between sentences. For short-text classification, the auxiliary context is unavailable during the prediction phase and thus, the sequence connection (through the separator tag [SEP]) between short text and auxiliary context is not present. This scenario commonly occurs in real-world applications when the auxiliary context is unavailable for classification (e.g., lack reviews when categorizing new Amazon products).

There are two popular approaches for integrating contextual information in short-text classification. One of the approaches [5] uses entity linking and conceptualization to link the phrases in the short text to concepts in a knowledge base. However, the performance of the short-text classifier is challenged by the entity linking task and the availability of a knowledge base of sufficiently high quality. Other approaches [35, 46] extract topic distributions from the short text and combine them with the short-text semantic information to create additional contextual features for classification. The main drawback of these approaches is that the model cannot adapt (zoom in or zoom out) to different topical granularities to create the necessary context. We overcome these challenges by generating necessary contextual information given the short text at different topic granularities and without explicitly linking the entities to knowledge bases.

In this work, we reformulate the short-text classification problem and design a new self-supervised learning framework with two related tasks, namely, the sequence auxiliary context generation task (which acts as our proxy task) and the class prediction task (which acts as the fine-tuning task). Note that both the related tasks, in a self-supervised learning paradigm, depend only on the auxiliary context that is systematically obtained from the data and do not rely on any additional manual labeling. The sequence generation problem utilizes short text to generate a conditional pseudo-auxiliary context distribution based on statistical inference from the training data distribution. The classification task predicts the final label based on an aggregation of short-text features and the auxiliary context distribution. Figure 1 illustrates the proposed learning framework. To achieve this, we propose the Pseudo-Auxiliary Context generation network for Short-text modeling (PACS). The model architecture consists of two mutually trained sub-networks, namely, the Context Generation Network (CGN) and Prediction Network (PN). CGN is a sub-network of the Self-Attentive Bi-directional Long Short-Term Memory (Bi-LSTM) network, which handles the sequence generation sub-problem using an encoder-decoder architecture that encodes short text and decodes to the corresponding auxiliary context. Additionally, to maintain consistency in the domain, we need the auxiliary context to be conditioned upon the short text. The sequential dependence of the decoder on the encoder retains the conditionality constraint for the auxiliary context generation. The PN utilizes these features for the final prediction task. The primary goal of PACS is to learn how to generate auxiliary context in a self-supervised manner for the task of short-text classification. The main contributions of the article are summarized as follows:

Fig. 1.

Fig. 1. The proposed self-supervised learning framework that generates auxiliary context for the problem of short-text classification. The dotted lines represent the data available during the training but inaccessible during prediction. Error correction is the gradient update to the Context Generation and Prediction Network using back-propagation.

  • We reformulate the short-text classification problem using self-supervised learning for leveraging auxiliary context through conditional sequence generation and a predictor to map features to the class label.

  • Inspired by the human focus mechanism, we develop PACS, a novel self-supervised learning architecture that consists of two sub-networks that are jointly trained on a final prediction task. A context generation network, which is a self-supervised model, first generates sequential auxiliary context distribution from the short text. The predictor network then applies these features to learn a model that maps features to their final class label.

  • We perform an extensive set of experiments across several real-world datasets to evaluate the performance of PACS against various state-of-the-art baseline methods. We also analyze the effectiveness of PACS and its sensitivity to dataset size and auxiliary context availability. Additionally, we qualitatively study the model to understand the reasons for its effectiveness.

The rest of this article is organized as follows: Section 2 discusses the relevant background for the problem. Section 3 reformulates the short-text classification problem using self-supervised learning and introduces the architecture of our proposed PACS model. Section 4 describes the real-world datasets, state-of-the-art baselines, and performance metrics used to evaluate the PACS model. We also show the performance along with the model interpretability study of the PACS’s architecture. Finally, Section 5 concludes our article.

Skip 2RELATED WORK Section

2 RELATED WORK

We discuss the background work related to two main sub-areas of research associated to the proposed model: short-text classification and attention mechanism in neural models.

2.1 Short-text Classification

One line of research in short-text classification relies on explicit feature construction using human-designed sparse features. Cavnar and Trenkle [4] employ \( n \)-gram features for text classification and in [29, 32, 36], the authors utilize more complex features such as POS tags and dependency parsing to improve the prediction. Another line of research works aim to leverage knowledge bases for enriching the information of short texts. In [10], Wikipedia is adopted to enrich the information retrieved from short text. Wang et al. [40] map the short text to a set of relevant concepts and leverage Probase to classify the obtained features. Explicit feature modeling generates human-interpretable representations but does not capture contextual semantic information. Another group of research works focus on probabilistic topic identification of short text. Latent Dirichlet Allocation (LDA) [3, 15] leverages word co-occurrence to represent short text as a distribution over its topics and vice versa. Additionally, Non-negative Matrix Factorization (NMF) has been successfully applied to short-text topic modeling [35].

Recently, implicit models have gained popularity due to the proliferation of deep learning algorithms. The models map the original text onto a semantic dense vector in a latent space. Word2Vec [26] and GloVe [31] provide word representations according to their context and co-occurrence, respectively. In [18], the authors utilize a combination of convolutional network to capture semantic features and recurrent network to obtain sequential features for short-text classification. Character-level convolutional neural network (CNN)-based models [7, 14, 48] provide semantic features from character \( n \)-grams using CNN filters for text classification. Bi-LSTM and Self-attention models [8, 20, 22, 44] encode short text and relations between the words for text classification. Topic-model based approaches [21, 43, 47, 50] leverage cross-text word co-occurrence-based topic modeling over large documents (or pseudo-documents) to learn word embeddings and further utilize them to solve the problem of data sparsity in short text and improve the classification performance. In [38], the authors utilize auxiliary signals in the area of e-commerce such as purchasing intent to construct graphs of similar short-text queries and products. A graph convolution network (GCN) [17] is applied to this graph for improved performance. The graph of similar items limits data sparsity by providing additional information for enhanced classification. Meng et al. [25] utilize weak supervision to generate pseudo-documents for hierarchical text classification. Although implicit models effectively capture syntactic and semantic information, they lack the ability to capture information from knowledge bases. Topic memory networks [46] and short-text classification with knowledge-powered attention [5] use both short text and encoded knowledge bases to enhance prediction. However, these models do not focus on relevant knowledge bases. Hence, to alleviate this problem, the proposed PACS model utilizes sample-based auxiliary context to capture features from both the short text and its domain.

2.2 Attention Mechanism in Neural Models

Attention mechanisms have been proven to be effective in various types of neural models. The mechanism can be broadly grouped into two categories (1) vanilla attention and (2) self-attention. Bahdanau et al. [2] applied the vanilla attention mechanism to compute the relevance score between query and input tokens for a machine translation task. Inspired from these works, we employ self-attention to understand the importance of individual words to sentences in PACS during the context generation phase.

Topic memory networks [28, 42, 45, 46, 49] employ pre-trained topic models on external data in order to encode latent topic representations for short-text classification. However, such an approach loses out on the domain relevance by dismantling correspondence between short text and auxiliary context and utilizing a bag of auxiliary context instead of individually aligned samples. Samples with corresponding auxiliary context maintain relevance and remain domain specific. Additionally, the topic modeling and short-text classification architectures are independent and hence, the topic features possess limited relevance to the short-text classification task. Also, classifier models that leverage knowledge bases have shown good performance [12, 23, 30]. The Short Text Classification with Knowledge-powered Attention (STCKA) [5] model jointly trains the topic modeling and classification task, but the scope of the knowledge base is wide. There is no mechanism to consider domain relevance for validation. We need a framework to model auxiliary context generation conditional on short text. This maintains the relevance of the sample’s features and leads to improved final prediction. The context generation needs to be contingent on the final class to learn features significant for the classification task.

Skip 3PACS MODEL ARCHITECTURE Section

3 PACS MODEL ARCHITECTURE

In this section, we introduce the overall self-supervised learning framework for short-text classification by leveraging the auxiliary context. We describe the overall architecture of PACS and its application in the context of short-text classification.

3.1 Problem Statement

Let \( D \) denote the full dataset out of which \( D_{T} \) and \( D_{V} \) denote the training and validation sets, respectively; i.e., \( D=D_T \cup D_V \). Each element \( \lbrace st_i, ac_i, y_i\rbrace \in D_{T} \) and \( \lbrace st_j,ac_j,y_j\rbrace \in D_{V} \) is the set of short text, its auxiliary context, and final class label, respectively. For \( D_{V} \), \( ac_j = \varnothing \) and \( y_j \in \lbrace y_1, y_2,\ldots , y_{|class|}\rbrace \) is the prediction variable. The primary goal of classification is to optimize a model \( P_{\theta } \) parameterized by \( \theta \), such that (1) \( \begin{equation} \theta = \underset{\theta }{\arg \min }\left(\sum _{i=1}^{|D_T|}-y_i\log \left(P_{\theta i}\right)\right). \end{equation} \)

Let \( X_p \) represent a probability distribution over parameters \( p \). Under the assumption that \( X_{st_i}\sim X_{st_j} \), we estimate parameter set \( \theta \) as a strong predictor for \( D_V \). In PACS, we leverage \( ac_i \in D_T \) to learn a robust \( \theta \) for prediction when \( ac_j \) is not available: (2) \( \begin{gather} \lambda = \underset{\lambda }{\arg \max }\left(\sum _{i=1}^{|D_T|}P\left(ac_{\lambda i}|st_i\right)\right) \ni \forall i = 1\rightarrow |D_T|, X_{\lambda i} \sim X_{ac_i} \end{gather} \) (3) \( \begin{gather} \theta = \underset{\theta }{\arg \min }\left(\sum _{i=1}^{|D_T|}-y_i\log \left(X_{st_i, ac_{\lambda i}}\right)\right), \end{gather} \) where \( \lambda \in \mathbb {R}^{k} \) (k is the number of parameters equivalent to weights in a neural network) is the set of sequence generator parameters estimated by maximizing the probability of generated auxiliary context \( ac_{\lambda i} \) given short text \( st_i \) over the training set \( i=1\text{ to} |D_T| \) while maintaining similarity between distributions \( X_{\lambda i} \) and \( X_{ac_i} \). \( \theta \in \mathbb {R}^{|class| \times 2k} \) is the set of classifier parameters that minimize the cross-entropy between target class \( y_i \) and combined features of short text and the generated auxiliary context \( X_{st_i,ac_\lambda } \).

3.2 PACS Model Architecture

Figure 2 illustrates the overall architecture of our framework. The model consists of two modules corresponding to the self-supervised learning pipelines: CGN and PN. The CGN learns a sequence generation function to map short text to its corresponding auxiliary context, and the PN learns a prediction model to classify a concatenation of encoded short text and auxiliary context to its sparse categories.

Fig. 2.

Fig. 2. The overall architecture of the proposed PACS model. During the training phase, the self-supervised Context Generation Network (CGN) encodes the input short text and decodes to the auxiliary context. For validation, the CGN network utilizes the short text to generate a distribution similar to the statistically inferred auxiliary context. The prediction network utilizes the short-text encoding and generated auxiliary context distribution for the final class prediction.

3.2.1 Context Generation Network (CGN).

The aim of the CGN is to generate the auxiliary context for a given short-text sequence. We model the problem as a sequence generation task, where the input short text \( \lbrace st_1, st_2,\ldots , st_{|D|}\rbrace \in ST \) generates the corresponding output auxiliary context \( \lbrace ac_1, ac_2,\ldots , ac_{|D|} \rbrace \in AC \). To achieve this goal, we use a self-supervised model based on self-attentive Bi-LSTM networks. More formally, we optimize for parameters of the generator model \( f_\lambda :X_{st}\rightarrow X_\theta \ni X_\theta \sim X_{ac} \), where \( X_{st}, X_\theta , X_{ac} \) is the distribution of short-text, generated, and original auxiliary context distribution, respectively.

The embedding layer converts the sparse one-hot \( ST \) into a dense \( k \)-dimensional representation \( \lbrace st_{i1}, st_{i2},\ldots , st_{ik} \in st_i\rbrace \) (\( k \) is decided empirically). The embedding layer has multiple variants: Word2Vec [26], BERT [8], and XLNet [44]. Word2Vec consists of pre-trained semantic vectors that use a word’s context from a large corpus. BERT and XLNet are language models trained on large corpora with word masking and word permutation, respectively. For BERT and XLNet, we extract the last layer for word embeddings.

The Bi-LSTM layer encodes the sequence with a forward \( h_{ft} \) and backward \( h_{bt} \) LSTM [37]. Back-propagation through time (BPTT) [41] simultaneously updates the hidden states of the forward (\( f_t \)) and backward (\( b_t \)) pass. The weights provide sequential information of the sentences. We adopt Rectified Linear Unit (ReLU) as our activation function for non-linearity and faster convergence: (4) \( \begin{gather} h_{f_t} = LSTM(w_{f_t},h_{f_{t-1}}) \end{gather} \) (5) \( \begin{gather} h_{b_t} = LSTM(w_{b_t},h_{b_{t-1}}) \end{gather} \) (6) \( \begin{gather} o_t = max\lbrace 0,W[f_t,b_t]x_t+b_t\rbrace , \end{gather} \) where \( h_{f_t} \) and \( h_{b_t} \) are the hidden LSTM units of the forward and backward pass, respectively, and \( o_t \) is the ReLU activation unit for the combined weights. To capture long-term sentence dependencies, we employ the self-attention network. It computes attention weights that denote the significance of relations between words in the sentence. Given that the maximum sequence length of input is \( n \), the weight matrix \( \alpha \in \mathbb {R}^{n \times h} \) for the dot-product attention over \( h \) hidden units and the final sentence encoding \( e_t \) scaled with the attention weights (\( \alpha _{ij} \)) are given by (7) \( \begin{equation} \alpha _{ij} = \frac{o_io_j^T}{\sqrt {2h}}\text{;}\quad e_t = \sum _{j}\frac{exp(\alpha _{tj})}{\sum _{t}exp(\alpha _{tj})}o_{tj}. \end{equation} \)

The attention weights (\( \alpha _{ij} \)) indicate the significance of the interaction between encoded sequential outputs \( o_i \) and \( o_j \) in the latent space to the final encoded output at the timestep \( e_t \). The encoding \( e_t \) initiates the decoder model. The decoder for the auxiliary context is similarly modeled with attention matrix \( \beta \in \mathbb {R}^{n \times h} \) and sequence decoding \( d_t \) given by the following equations: (8) \( \begin{gather} \beta _{ij} = \frac{o_io_j^T}{\sqrt {2h}}\text{;}\quad d_t = \sum _{j}\frac{exp(\beta _{tj})}{\sum _{t}exp(\beta _{tj})}o_{tj}. \end{gather} \)

However, this problem encounters an information bias. The short text contains little evidence to support a full reconstruction of the auxiliary context (\( X_{ac} \)). However, for our primary problem of classification, we merely expect additional information from the auxiliary context’s space to support the short text. Hence, the CGN will only predict a distribution \( X_\theta \sim X_{ac} \) for the additional information. To this end, we employ KL-divergence as our loss function \( L_{KLD} \): (9) \( \begin{equation} L_{KLD}(X_{\theta },X_{ac}) = X_{\theta } \log \left(\frac{X_{\theta }}{X_{ac}}\right), \end{equation} \) where \( X_{\theta } \) and \( X_{ac} \) are the distributions that need to be compared. Unlike other prominent loss functions designed to evaluate point-wise categorical predictions, KL-divergence measures the similarity between prediction distributions and the ground truth. The significance of \( X_\theta \) to \( y \) is successfully demonstrated in our experimental results presented in Section 4.3.

3.2.2 Prediction Network (PN).

In the PN, we utilize a dense network to predict the final class label \( y \) from features \( x_i \in x \) (\( \forall ~i \)) extracted from a concatenation of input text \( e_i \) and the auxiliary context’s predicted distribution \( d_i \): (10) \( \begin{equation} x_i = e_i \odot d_i. \end{equation} \)

The final probability for each category \( y_i \in y \) is given by (11) \( \begin{equation} P(y_i|x_i) = \frac{exp\left(\sum _{j=1}^{2k}w_{ij}x_{ij}+b_i\right)}{\sum _{i=1}^{c}exp\left(\sum _{j=1}^{2k}w_{ij}x_{ij}+b_i\right)}, \end{equation} \) where \( c \) is the number of classes, \( k \) is the number of output units in the short-text encoder and auxiliary context decoder, \( w, x \in \mathbb {R}^{c \times 2k} \), \( y \in \mathbb {R}^c \), and \( x_i = \lbrace x_{i1}, x_{i2},\ldots , x_{i2k}\rbrace \). We utilize cross-entropy as the loss because the prediction class labels are categorical: (12) \( \begin{equation} L_{CE}(\hat{y},y) = -\sum _{i=1}^{c}\hat{y_i}\log (y_i). \end{equation} \)

3.2.3 PACS Algorithm.

Algorithm 1 provides a high-level pseudo-code of the proposed PACS model. The goal is to estimate the generator model \( f_\lambda \) and predictor \( P_\theta \) given input training set \( D_T \). Lines 4 and 5 encode the short text and generate auxiliary context distribution, respectively. The losses for CGN (\( l_{cgn} \)) and Prediction Network (\( l_{pn} \)) are calculated from lines 6 to 9. Based on the losses, \( \lambda \) and \( \theta \) are updated in line 11 through back-propagation. The updated \( \lambda \) and \( \theta \) form the parameters for our generator model \( f \) and predictor model \( P \), respectively. The semantic representations in PACS are empirically set to 300 dimensions (\( k \) = 300). We adopt ReLU as our activation unit to introduce non-linearity and Dropout (\( p \) = 0.5) to avoid over-fitting on the training set.

Skip 4EXPERIMENTAL SETUP Section

4 EXPERIMENTAL SETUP

In this section, we describe the datasets in our experiments and then provide baselines along with implementation details.

4.1 Dataset Description

To analyze the performance of the proposed PACS model and compare it against other state-of-the-art text classification approaches, we conducted comprehensive experiments using several real-world datasets. Table 2 summarizes additional details including some basic statistics of the datasets.

Table 2.
DatasetAvg LenMax LenNo. of# Samples
STACSTACclassesTrainVal
Amazon970524,88143912,000608,000
HuffPost9204424341120,55180,341
RT3126252,4731217,88611,924
ArXiv8150335584124,60016,400
  • ST and AC columns represent the number of Short Text and Auxiliary Context samples, respectively.

Table 2. Dataset Statistics (Average Sentence Length, Maximum Sentence Length, Number of Classes, and Number of Training/Validation Samples in Each of the Datasets)

  • ST and AC columns represent the number of Short Text and Auxiliary Context samples, respectively.

  • Amazon Reviews2: The dataset contains 142.5 million Amazon reviews and metadata of 43 product categories from May 1996 to July 2014. In our experiments, we utilize the “product title” as the short text to predict the “product category.” During training, we employ the corresponding “review body” as the auxiliary context. We randomly sample a uniform number of data points from each class for our experiments. We also study the effect of this sampling and the scalability of our model in Section 4.2.

  • HuffPost News3: The dataset consists of 200K news headlines and metadata of 41 categories from 2012 to 2018. We utilize the “headline” as the short text to predict the “category.” For training, we leverage the corresponding “short_description” as the auxiliary context.

  • Rotten Tomatoes (RT) Movies4: It contains 30K movies and metadata of 12 genres from 1914 to 2018. The experiments use the “Title” as the short text to predict “Genre.” The training employs the corresponding “Description” as the auxiliary context.

  • ArXiv Papers5: This dataset contains 41K research papers from arXiv labeled with 42 tags/categories. In the experiments, we utilize “title” as the short text to predict the category “term” and generate the corresponding auxiliary context “summary.”

To study the performance of PACS on a varying number of classes, we consider \( n \)-class versions of our datasets. An \( n \)-class version of the dataset only utilizes the top \( n \) classes that have the most number of samples for training and testing phases of PACS and the baselines. We did not include other text classification benchmark datasets (such as the GLUE [39]) because they lack auxiliary context. In such cases, our model will mirror the performance of its embedding layer (Word2Vec, BERT, or XLNET) as there is no enrichment of features from the auxiliary context. In addition to this, although the datasets contain an auxiliary context, they are not considered \( (AC=\varnothing) \) in the evaluation procedures of our experiments. This simulates the real-world problem where the context is available for learning but unavailable for classification of new samples.

4.2 Performance Comparison

We perform several experiments to compare our model against other state-of-the-art methods. Additionally, we compare the performance of PACS by varying the number of classes, data points, and the availability of auxiliary context.

4.2.1 Comparison with Baselines.

We compare PACS with several state-of-the-art baseline models for short-text classification. In the training phase, all the models have access to both the short text and auxiliary context. For the validation phase, the models exclusively use the short text for class prediction. The baseline models used for comparison in our experiments are as follows:

  • Latent Dirichlet Allocation (LDA)6 [3]: In our experiment, we learn topic models \( T \) as a distribution over the combination of short text and auxiliary context (\( T_\theta \sim ST \cup AC \)). Simultaneously, the final topic matrix allows us to extract word vectors as a distribution over the topic models.

  • Topic Memory Networks (TMNs) [46]: The model is a combination of three major units: neural topic model to infer latent topics, topic memory mechanism to extract features from latent topics, and a final prediction model to map features to a class.

  • Short Text Classification with Knowledge-powered Attention (STCKA)7 [5]: The model learns the Concept to Short Text (C-ST) and Concept to Concept (C-CS) attention from auxiliary context. This conceptual knowledge is applied to predict short text’s classes.

  • Bi-directional Transformers (BERT) [8]: The model utilizes transformers to capture the co-dependence of different sentence units as attention weights. BERT achieves this through training a language model by masking certain inputs. We adopt the large pre-trained BERT model and fine-tune it on our datasets. The fine-tuning inputs are a concatenation of short text, a separator label [SEP], and the auxiliary context. Also, we adopt two variations of fine-tuning; one is trained only on short text (BERT-ST) and other has additional access to auxiliary context (BERT-STAC).

  • Auto-regressive Transformers-XL (XLNet) [44]: Unlike BERT, XLNet learns the language model through maximizing likelihood over all permutations for input masks. We adopt the large pre-trained XLNet model and fine-tune it on our datasets. Similar to the variants of BERT mentioned above, we use two variations based on training phase: XLNet-ST and XLNet-STAC.

4.2.2 Implementation Details.

LDA provides vectors for words in the dataset. The sentence vector is an average over the word vectors. The sentence vector trains the dense classifier to optimize for the final class label. The remaining models are composed of end-to-end frameworks that output the class labels through intermediate feature extraction. To maintain fair comparison, we consistently set embedding dimensions (\( k=300 \)) and the number of hidden units (\( h=128 \)) to be constant across models. We empirically tune the other hyper-parameters, within our computational capacity for best results. For sequential models involving LSTMs, we use BPTT to speed up the training process. We train all the networks including PACS with mini-batch Adam Optimizer [16] with a batch size of 64. We use gensim [34] for the baseline LDA model. For BERT and XLNet, we use their keras implementations.8 The developers for these modules provide a precise mechanism for integrating classification. For TMN, we adopt the code provided by the authors [46]. For STCKA and PACS, we build on the layers in Tensorflow 2.0 [1] to model the architectures. We utilize K-fold cross-validation with a train, validation, and test ratio of 75:10:15 in our experimental setup. For training PACS, we use an NVIDIA P40 with 12GB of VRAM. We implemented PACS with keras Bi-LSTM and Attention layers in congruence with its Merge layer for joint training. PACS is trained with a contrastive learning procedure with one negative sample for each positive sample for the different classes. Contrastive learning ensures better discriminative power in the classification procedure as the model is able to differentiate between positive and negative samples for each class. This also reduces the class bias by maintaining a constant ratio of positive and negative samples for each class. The final hyper-parameters for PACS and our baselines are summarized in Table 3. Note that PACS requires considerably fewer parameters compared to XLNet and BERT, thus decreasing its reliance on availability of computational resources (high GPU memory).9

Table 3.
ModelsDropoutDenseMax Seq# Model
RateUnitsLengthParameters
LDANA128NA38.4 K
TMN0.5641008.98 M
STCKA0.5641001.45 M
BERT0.264100110.3 M
XLNet0.264100146.8 M
PACS0.5641005.48 M

Table 3. Hyper-parameter Values in Our Experiments for PACS and the Baselines

4.2.3 Performance Comparison Results.

We validate the models’ predictions using two evaluation metrics: Accuracy (estimates total prediction) and class-weighted F1-score (estimates the harmonic mean between precision and recall). The metrics are computed for each class and the reported results are averages weighted by the number of samples in each class. Table 4 depicts the results of our experiments. We observe that our model outperforms the current state of the art by \( \approx \hspace{-2.3pt}8\% \) in Accuracy and \( \approx \hspace{-2.3pt}7.5\% \) in F1-score for fewer number of classes, and by \( \approx \hspace{-2.3pt}200\% \) in Accuracy and \( \approx \hspace{-2.3pt}45\% \) in F1-score for higher number of classes. We conjecture that these improvements are primarily due to the additional contextual features in the auxiliary text distribution that are generated by our CGN module in PACS.

Table 4.
  • ST and STAC report the values when the models are trained without and with additional auxiliary context, respectively. PACS (ours) reports the results for PACS with BERT as the primary embedding layer. The Improv row compares the performance improvement of PACS (ours) relative to XLNet (STAC).

Table 4. Performance Comparison of the Proposed PACS Model with Several Baseline Methods across Various Datasets and Evaluation Metrics: (a) Accuracy and (b) F-score

  • ST and STAC report the values when the models are trained without and with additional auxiliary context, respectively. PACS (ours) reports the results for PACS with BERT as the primary embedding layer. The Improv row compares the performance improvement of PACS (ours) relative to XLNet (STAC).

4.2.4 Sensitivity to Dataset Attributes.

In these experiments, we analyze the reliability of PACS on different dataset attributes. For our study, we vary the datasets in the number of data points and availability of auxiliary data.

  • Number of data points: We randomly sample subsets of various size from the datasets and measure the performance metrics and computational time.

  • Auxiliary context: We remove the auxiliary context and only utilize the short text to analyze the change in performance.

Figure 3 shows the performance variation according to the dataset attributes. In this experiment, we notice a significant increase of 25% to 106% in Accuracy as the number of data points increases. This indicates better generalizability of the auxiliary context generated by the CGN module with an increase in data points. Also, absence of auxiliary context hinders the model’s performance and decreases Accuracy by 2% to 56%. These results clearly illustrate the utility of auxiliary context generation. This conclusion is further supported by the qualitative evidence presented in Section 4.3.3.

Fig. 3.

Fig. 3. Sensitivity to dataset attributes (best viewed in color). (a) The graph depicts the shift in F-score with varying dataset sizes: 5,000, 10,000, 20,000, and 50,000 samples. (b) The graph shows the change in F-score with and without Auxiliary context.

4.3 Model Interpretability

In these sets of experiments, we qualitatively analyze the influence of generated auxiliary context to the overall prediction and study the effect of different sentence segments to the generation process. This helps us in understanding PACS’s attention mechanism and better interpreting its underlying activations.

4.3.1 Significance of Auxiliary Context.

Auxiliary context provides additional information to improve performance of the prediction network. We understand the context’s exact contribution through the weights learned by the dense classifier. Figure 4 shows a heat map of the classifier weights. We notice that weights in the case of XLNet-STAC are significantly focused on the short text. This indicates a lack of attention toward the auxiliary context. This is resolved in PACS, which uniformly captures features from both short text and auxiliary context.

Fig. 4.

Fig. 4. Significance of auxiliary context. The top and bottom four rows represent the weights of short text and auxiliary context, respectively. (a) and (b) present the activation of the PN’s hidden units for 10 classes in Amazon Reviews with XLNet-STAC and PACS (XLNet), respectively. Darker cells represent more relative weightage. The weights for short text look paler in PACS because of normalization.

4.3.2 Qualitative Significance of Auxiliary Context.

In this experiment, we qualitatively analyze the top-ranked words in the auxiliary context that aid overall prediction improvement. For each sample short text, we analyze the attention weights and pick words with maximum weights toward the final prediction (shown in Table 5). We observe that auxiliary context supports the prediction of short text by identifying significant words that add semantic relevance to the ambiguous short text. As we observe from Table 5, this enables estimation of a better prediction model.

Table 5.

Table 5. Qualitative Results Showing the Significance of Auxiliary Context

4.3.3 Attention Weights for Prediction.

We analyze the segments of text that help in the sequence encoding and decoding in CGN. We run PACS on a sample test case and analyze the attention weights given to each word embedding during the sequence generation from short text to auxiliary context.

Figure 5 displays the attention weights for a sample sequence. We can observe that semantically rich words (e.g., mouse, keyboard) receive more attention (higher weights) than stop words (e.g., and, the) for final prediction. Stop words do not provide discriminative features and thus have limited utility in the problems of classification. PACS enriches the feature set by ignoring stop words and focusing on semantically significant words through its attention modules. This improves discriminative classification and thus, we observe a corresponding increase in performance.

Fig. 5.

Fig. 5. Attention weights of sequence generation from short text to auxiliary context (best viewed in color). Figure illustrates activations for an example from the Amazon product catalog.

4.4 Ablation Study

We study the importance of various embedding layers and the presence of the self-attention mechanism and Bi-LSTMs in this section.

4.4.1 Embedding Layers Study.

In this experiment, we replace the embedding layer in PACS and study the difference in performance.

  • Word2Vec (W2V): We train a skip-gram model [26] on the dataset and extract the word-level features for the CGN encoder and decoder.

  • Pre-trained models: We extract the features from the last layer of GPT, BERT, BART, and XLNet language models and utilize them in CGN.

Table 6 provides the results for variants of PACS on the Amazon Reviews dataset. These results demonstrate that variation in the embedding layer leads to insignificant change in performance. Hence, PACS does not rely on the quality of representations but only leverages the layer for a dense dimensionality reduction from a one-hot sparse semantic space.

Table 6.
MetricsAccuracyF-score
# Classes3104331043
PACS (W2V)85.459.731.24.811.553.300
PACS (GPT)85.559.831.21.813.553.302
PACS (BERT)85.459.831.23.812.556.301
PACS (BART)85.659.931.20.811.556.301
PACS (XLNet)85.759.631.70.814.555.305
  • Experiment to show the importance of the embedding layer. The embedding layers tested are W2V, GPT, BERT, BART, and XLNet.

Table 6. Significance of the Embedding Layer

  • Experiment to show the importance of the embedding layer. The embedding layers tested are W2V, GPT, BERT, BART, and XLNet.

4.4.2 Self-attention and Bi-LSTM Study.

In this experiment, we study the individual contribution of the self-attention mechanism and Bi-LSTM component to the overall architecture of PACS. We test two variants of PACS, namely, one without self-attention (w/o SA) and one without Bi-LSTM (w/o BL). We test these variants against PACS on each of the datasets for short-text classification. We perform the ablation by blocking weight updates to the ablated component during the training phase. Hence, the number of parameters and dimensions remains intact and fairly comparable. Additionally, we also train other model variants without any auxiliary context (w/o AC) and without any short-text information (w/o ST) to analyze the significance of auxiliary context and short text, respectively, to the overall performance of the model.

Table 7 shows the results of the ablation study. In this experiment, we can observe that SA, Bi-LSTM, AC, and ST contribute 8% to 30%, 2% to 9%, 13% to 46%, and 10% to 32% to the overall performance accuracy, respectively. The self-attention mechanism captures the contribtuion of the inter-token dependence to the overall representation, whereas Bi-LSTMs model the sequential/ positional information of the vectors. The lack of Bi-LSTM does not cause significant changes except in a high number of classes. Also, we observe that auxiliary context plays the most prominent role in enhancing the prediction followed by SA. Thus, we conclude that auxiliary context generation and inter-token relations play a more vital role than sequential information.

Table 7.
  • Performance comparison of the contributions from different components of the model such as Self Attention (SA), Bi-LSTM (BL), Auxiliary Context (AC), and Short Text (ST) to the overall performance. PACS (pipe) is the pipeline version of PACS that independently learns the short-text and auxiliary context features.

Table 7. Ablation Study Results on (a) Accuracy and (b) F-score Measures

  • Performance comparison of the contributions from different components of the model such as Self Attention (SA), Bi-LSTM (BL), Auxiliary Context (AC), and Short Text (ST) to the overall performance. PACS (pipe) is the pipeline version of PACS that independently learns the short-text and auxiliary context features.

4.4.3 Pipeline vs. Joint-learning Study.

We conduct this experiment to analyze the difference between separately (pipeline) and jointly learning the features for short text and the corresponding auxiliary context. For this, we independently learn the features of short text with XLNet and train the CGN network to generate auxiliary context. We concatenate these independently learned short-text and auxiliary context features and input them to train a dense prediction network that outputs the final class label. The experiment is conducted on the same cross-validation split as the main PACS model.

From the results in Table 7, we observe that the joint-learning model is able to significantly outperform the pipeline model. The reason for these improvements is the direct connection between the CGN and PN module that allows the network to capture the class signal and backpropagate it to optimize task-specific auxiliary context generation and feature extraction. Thus, we can conclude that jointly learning CGN is better for enhancing the features of short text.

4.4.4 Loss Function Study.

We conduct this experiment to analyze the difference in PACS’s performance based on different loss functions. For this study, we consider the standard loss functions for classification, cross-entropy [9, 24], InfoNCE [27], and hinge loss [11], and compare it to our choice, the KL-divergence loss. The experiment is conducted on the same cross-validation split as the main PACS model.

From the results in Table 8, we observe that the KL-divergence loss outperforms the other loss functions in the comparative study. Thus, we can conclude that KL-divergence loss is better for generating pseudo-auxiliary context and improving classification performance.

Table 8.
MetricsAccuracyF-score
# Classes3104331043
PACS (Cross Entropy)75.048.920.8.707.445.195
PACS (InfoNCE)82.255.928.0.778.519.269
PACS (Hinge Loss)78.652.324.3.741.4830.232
PACS (KL-divergence)85.759.631.7.814.555.305
  • Performance comparison using different loss functions. The loss functions tested are cross entropy, InfoNCE, hinge loss, and KL-divergence

Table 8. Analysis for the Significance of the Loss Function

  • Performance comparison using different loss functions. The loss functions tested are cross entropy, InfoNCE, hinge loss, and KL-divergence

Skip 5CONCLUSION Section

5 CONCLUSION

We reformulated the short-text classification as a self-supervised learning problem that leverages the sample’s auxiliary context through conditional sequence generation. Furthermore, a predictor network then utilizes the short-text encoding and the generated sequence as features for the final class prediction. We developed the Pseudo-Auxiliary Context generation network for Short-text modeling (PACS) to employ the reformulation and comprehensively leverage the auxiliary context for the problem of short-text classification. The network consists of two sub-modules: context generation network and prediction network. PACS jointly trains the sub-modules in an end-to-end self-supervised learning framework to exclusively capture features relevant to the class prediction. We evaluated PACS through comparative studies against state-of-the-art baselines on benchmark datasets. Our experiments indicate that PACS outperforms several baselines on short-text classification using popular metrics such as accuracy and F-score. Additionally, we also performed qualitative interpretability analysis to identify the function of inner mechanisms for some sample cases. Through the trained weights, we observed that the positive contribution of auxiliary context increases with the number of classes. The attention weights demonstrate the effectiveness of context generation. Furthermore, we also performed an ablation study to comprehend the contribution of individual networks to the overall architecture.

APPENDICES

A HYPER-PARAMETER TUNING

To support reproducibility, this section discusses the sensitivity of PACS on the initial hyper-parameters. Furthermore, we provide the details of hyper-parameters utilized in our final experimental setup. We vary the following hyper-parameters and analyze their impact on the model performance and loss.

Skip A.1Activation Function Section

A.1 Activation Function

We evaluate three activation functions: Sigmoid, Tanh, and ReLU. Qu et al. [33] show that ReLU and Tanh are better than sigmoid for deep models. Figure 6(a) presents a similar finding for our PACS model. ReLU performs slightly better than Tanh, possibly because it induces sparsity for least significant values.

Fig. 6.

Fig. 6. Sensitivity of PACS to hyper-parameters: (a) activation function and (b) dropout probability.

Skip A.2Dropout Values Section

A.2 Dropout Values

Figure 6(b) demonstrates PACS’s loss on five dropout probabilities: 0.1, 0.2, 0.3, 0.4, and 0.5. We observe that dropout does not significantly affect the model’s loss. Hence, we choose 0.5 because it reduces the number of parameter updates.

Skip A.3Embedding Size Section

A.3 Embedding Size

Figure 7(a) demonstrates that increasing the embedding improves the model’s convergence loss. The reason is that additional parameters capture better features for words’ reconstruction. Also, it reveals that decreasing \( k \) may decrease the loss further. However, due to computational restrictions, we utilize the maximum possible \( k= \hbox{1,000} \). We believe, given higher memory, that the performance can improve for higher values of \( k \).

Fig. 7.

Fig. 7. Sensitivity of PACS to hyper-parameters: (a) embedding layer size, (b) dense units, and (c) maximum sequence length.

Skip A.4Dense Units Section

A.4 Dense Units

Figure 7(b) displays that increasing the dense units from 32 to 64 decreases the model’s convergence loss. However, increasing the number further to 128 and 256 results in an increase in the convergence loss. This is because the increase in number of parameters requires more epochs to optimize for the minimum. However, computational restraints inhibit our ability to spend more epochs.

Skip A.5Maximum Sequence Length Section

A.5 Maximum Sequence Length

Figure 7(c) depicts that increasing the sequence length from 25 to 100 improves the model’s convergence loss. However, increasing it further captures redundant information and leads to a decrease in performance. Hence, we set it to 100 in our experimental setup. Based on the above empirical studies, the final hyper-parameters in our experimental setup are given in Table 3.

B TRAINING PHASE

Figure 8 illustrates the training and validation loss over epochs during the training phase. We observe that the model converges in \( \approx \hspace{-2.3pt}40 \) epochs. The loss function is categorical cross-entropy.

Fig. 8.

Fig. 8. Epochs vs. Training Loss and Validation Loss.

Footnotes

  1. 1 https://www.scientificamerican.com/article/wired-for-categorization/.

    Footnote
  2. 2 https://s3.amazonaws.com/amazon-reviews-pds/tsv/index.txt.

    Footnote
  3. 3 https://www.kaggle.com/rmisra/news-category-dataset.

    Footnote
  4. 4 https://www.kaggle.com/ayushkalla1/rotten-tomatoes-movie-database.

    Footnote
  5. 5 https://www.kaggle.com/neelshah18/arxivdataset.

    Footnote
  6. 6 LDA models the topic distribution that we utilize as the text vectors of our baselines.

    Footnote
  7. 7 Due to unavailability of original code, we implemented STCKA based on the original paper and fine-tuned the hyper-parameters to obtain the best possible result.

    Footnote
  8. 8 https://github.com/CyberZHG/keras-bert , https://github.com/CyberZHG/keras-xlnet.

    Footnote
  9. 9 Code for the PACS model is shared on https://github.com/Akirato/PACS.

    Footnote

REFERENCES

  1. [1] Abadi Martín, Barham Paul, Chen Jianmin, Chen Zhifeng, Davis Andy, Dean Jeffrey, Devin Matthieu, Ghemawat Sanjay, Irving Geoffrey, Isard Michael, Kudlur Manjunath, Levenberg Josh, Monga Rajat, Moore Sherry, Murray Derek G., Steiner Benoit, Tucker Paul, Vasudevan Vijay, Warden Pete, Wicke Martin, Yu Yuan, and Zheng Xiaoqiang. 2016. TensorFlow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Conference on Operating Systems Design and Implementation (OSDI’16). USENIX Association, 265283.Google ScholarGoogle Scholar
  2. [2] Bahdanau Dzmitry, Cho Kyunghyun, and Bengio Yoshua. 2015. Neural machine translation by jointly learning to align and translate. In 3rd International Conference on Learning Representations (ICLR’15), Conference Track Proceedings, Bengio Yoshua and LeCun Yann (Eds.). http://arxiv.org/abs/1409.0473.Google ScholarGoogle Scholar
  3. [3] Blei David M., Ng Andrew Y., and Jordan Michael I.. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research 3, (Jan.2003), 9931022.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. [4] Cavnar William B. and Trenkle John M.. 1994. N-gram-based text categorization. In Proceedings of 3rd Annual Symposium on Document Analysis and Information Retrieval (SDAIR’94), Vol. 161175. Citeseer, 161175.Google ScholarGoogle Scholar
  5. [5] Chen Jindong, Hu Yizhou, Liu Jingping, Xiao Yanghua, and Jiang Haiyun. 2019. Deep short text classification with knowledge powered attention. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 62526259.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. [6] Choudhary Nurendra, Rao Nikhil, Katariya Sumeet, Subbian Karthik, and Reddy Chandan K.. 2022. ANTHEM: Attentive hyperbolic entity model for product search. In The 15th ACM International Conference on Web Search and Data Mining (WSDM’22). Association for Computing Machinery, New York, NY.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] Choudhary Nurendra, Singh Rajat, Bindlish Ishita, and Shrivastava Manish. 2018. Neural network architecture for credibility assessment of textual claims. arXiv preprint arXiv:1803.10547 (2018).Google ScholarGoogle Scholar
  8. [8] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT’19), Volume 1 (Long and Short Papers), Burstein Jill, Doran Christy, and Solorio Thamar (Eds.). Association for Computational Linguistics, 41714186. Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Dikshit Abhirup and Pradhan Biswajeet. 2021. Interpretable and explainable AI (XAI) model for spatial drought prediction. Science of the Total Environment 801 (2021), 149797. Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Evgeniy Gabrilovich and Shaul Markovitch. 2007. Computing semantic relatedness using Wikipedia-based explicit semantic analysis. In IJcAI, Vol. 7. 1606–1611.Google ScholarGoogle Scholar
  11. [11] Gentile Claudio and Warmuth Manfred K. K.. 1998. Linear hinge loss and average margin. Advances in Neural Information Processing Systems 11 (1998), 225231.Google ScholarGoogle Scholar
  12. [12] Ginsberg Allen, Weiss Sholom M., and Politakis Peter. 1988. Automatic knowledge base refinement for classification systems. Artificial Intelligence 35, 2 (1988), 197.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. [13] Hu Jian, Wang Gang, Lochovsky Fred, Sun Jian-tao, and Chen Zheng. 2009. Understanding user’s query intent with wikipedia. In Proceedings of the 18th International Conference on World Wide Web. 471480.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Huang Po-Sen, He Xiaodong, Gao Jianfeng, Deng Li, Acero Alex, and Heck Larry. 2013. Learning deep structured semantic models for web search using clickthrough data. In Proceedings of the 22nd ACM International Conference on Information & Knowledge Management. 23332338.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. [15] Ioffe Sergey. 2006. Probabilistic linear discriminant analysis. In European Conference on Computer Vision. Springer, 531542.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. [16] Kingma Diederik P. and Ba Jimmy. 2015. Adam: A method for stochastic optimization. In 3rd International Conference on Learning Representations (ICLR’15), Conference Track Proceedings, Bengio Yoshua and LeCun Yann (Eds.). http://arxiv.org/abs/1412.6980.Google ScholarGoogle Scholar
  17. [17] Kipf Thomas N. and Welling Max. 2017. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations (ICLR’17).Google ScholarGoogle Scholar
  18. [18] Lai Siwei, Xu Liheng, Liu Kang, and Zhao Jun. 2015. Recurrent convolutional neural networks for text classification. In 29th AAAI Conference.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Lee Ji Young and Dernoncourt Franck. 2016. Sequential short-text classification with recurrent and convolutional neural networks. In The 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL HLT’16), Knight Kevin, Nenkova Ani, and Rambow Owen (Eds.). Association for Computational Linguistics, 515520. Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Lewis Mike, Liu Yinhan, Goyal Naman, Ghazvininejad Marjan, Mohamed Abdelrahman, Levy Omer, Stoyanov Veselin, and Zettlemoyer Luke. 2020. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 78717880. Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Li Chenliang, Duan Yu, Wang Haoran, Zhang Zhiqian, Sun Aixin, and Ma Zongyang. 2017. Enhancing topic modeling for short texts with auxiliary word embeddings. ACM Transactions on Information Systems 36, 2, Article 11 (Aug. 2017), 30 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. [22] Lin Zhouhan, Feng Minwei, Santos Cícero Nogueira dos, Yu Mo, Xiang Bing, Zhou Bowen, and Bengio Yoshua. 2017. A structured self-attentive sentence embedding. In 5th International Conference on Learning Representations (ICLR’17), Conference Track Proceedings. OpenReview.net. https://openreview.net/forum?id=BJC_jUqxe.Google ScholarGoogle Scholar
  23. [23] Gregor Robert Mac. 1991. The evolving technology of classification-based knowledge representation systems. In Principles of Semantic Networks. Elsevier, 385400.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Mannor Shie, Peleg Dori, and Rubinstein Reuven. 2005. The cross entropy method for classification. In Proceedings of the 22nd International Conference on Machine Learning (ICML’05). Association for Computing Machinery, New York, NY, 561568. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. [25] Meng Yu, Shen Jiaming, Zhang Chao, and Han Jiawei. 2019. Weakly-supervised hierarchical text classification. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 68266833.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Mikolov Tomas, Sutskever Ilya, Chen Kai, Corrado Greg S., and Dean Jeff. 2013. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems. 31113119.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. [27] Oord Aaron van den, Li Yazhe, and Vinyals Oriol. 2018. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018).Google ScholarGoogle Scholar
  28. [28] Palangi Hamid, Deng Li, Shen Yelong, Gao Jianfeng, He Xiaodong, Chen Jianshu, Song Xinying, and Ward Rabab. 2016. Deep sentence embedding using long short-term memory networks: Analysis and application to information retrieval. IEEE/ACM Transactions on Audio, Speech, and Language Processing 24, 4 (2016), 694707.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. [29] Pang Bo, Lee Lillian, and Vaithyanathan Shivakumar. 2002. Thumbs up? Sentiment classification using machine learning techniques. In Proceedings of the ACL-02 Conference on Empirical Methods in Natural Language Processing-Volume 10. Association for Computational Linguistics, 7986.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. [30] Pechenizkiy Mykola, Puuronen Seppo, and Tsymbal Alexey. 2003. Feature extraction for classification in knowledge discovery systems. In International Conference on Knowledge-Based and Intelligent Information and Engineering Systems. Springer, 526532.Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Pennington Jeffrey, Socher Richard, and Manning Christopher D.. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP’14). 15321543.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Post Matt and Bergsma Shane. 2013. Explicit and implicit syntactic features for text classification. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 866872.Google ScholarGoogle Scholar
  33. [33] Qu Yanru, Cai Han, Ren Kan, Zhang Weinan, Yu Yong, Wen Ying, and Wang Jun. 2016. Product-based neural networks for user response prediction. In 2016 IEEE 16th International Conference on Data Mining (ICDM’16). IEEE, 11491154.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Řehůřek Radim and Sojka Petr. 2010. Software framework for topic modelling with large corpora. In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks. ELRA, Valletta, Malta, 4550.Google ScholarGoogle Scholar
  35. [35] Shi Tian, Kang Kyeongpil, Choo Jaegul, and Reddy Chandan K. 2018. Short-text topic modeling via non-negative matrix factorization enriched with local word-context correlations. In Proceedings of the 2018 World Wide Web Conference. 11051114.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Socher Richard, Huval Brody, Manning Christopher D., and Ng Andrew Y.. 2012. Semantic compositionality through recursive matrix-vector spaces. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. Association for Computational Linguistics, 12011211.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Sundermeyer Martin, Schlüter Ralf, and Ney Hermann. 2012. LSTM neural networks for language modeling. In 13th Annual Conference of the International Speech Communication Association.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Tayal Kshitij, Rao Nikhil, Agarwal Saurabh, Jia Xiaowei, Subbian Karthik, and Kumar Vipin. 2020. Regularized graph convolutional networks for short text classification. In Proceedings of the 28th International Conference on Computational Linguistics: Industry Track. International Committee on Computational Linguistics, Online, 236242. Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Wang Alex, Singh Amanpreet, Michael Julian, Hill Felix, Levy Omer, and Bowman Samuel R.. 2019. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In The Proceedings of ICLR.Google ScholarGoogle Scholar
  40. [40] Wang Fang, Wang Zhongyuan, Li Zhoujun, and Wen Ji-Rong. 2014. Concept-based short text classification and ranking. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management. 10691078.Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. [41] Werbos Paul J.. 1990. Backpropagation through time: What it does and how to do it. Proceedings of the IEEE 78, 10 (1990), 15501560.Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Weston Jason, Chopra Sumit, and Bordes Antoine. 2015. Memory networks. In 3rd International Conference on Learning Representations (ICLR’15), Conference Track Proceedings, Bengio Yoshua and LeCun Yann (Eds.). http://arxiv.org/abs/1410.3916.Google ScholarGoogle Scholar
  43. [43] Yang Yi, Wang Hongan, Zhu Jiaqi, Wu Yunkun, Jiang Kailong, Guo Wenli, and Shi Wandong. 2020. Dataless short text classification based on biterm topic model and word embeddings. In Proceedings of the 29th International Joint Conference on Artificial Intelligence (IJCAI’20), Bessiere Christian (Ed.). International Joint Conferences on Artificial Intelligence Organization, 39693975. Main track.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Yang Zhilin, Dai Zihang, Yang Yiming, Carbonell Jaime, Salakhutdinov Russ R., and Le Quoc V.. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. In Advances in Neural Information Processing Systems. 57545764.Google ScholarGoogle Scholar
  45. [45] Yang Zichao, Yang Diyi, Dyer Chris, He Xiaodong, Smola Alex, and Hovy Eduard. 2016. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 14801489.Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Zeng Jichuan, Li Jing, Song Yan, Gao Cuiyun, Lyu Michael R., and King Irwin. 2018. Topic memory networks for short text classification. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Riloff Ellen, Chiang David, Hockenmaier Julia, and Tsujii Jun’ichi (Eds.). Association for Computational Linguistics, 31203131. Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Zhang Lu, Ding Jiandong, Xu Yi, Liu Yingyao, and Zhou Shuigeng. 2021. Weakly-supervised text classification based on keyword graph. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. 28032813.Google ScholarGoogle ScholarCross RefCross Ref
  48. [48] Zhang Xiang, Zhao Junbo, and LeCun Yann. 2015. Character-level convolutional networks for text classification. In Advances in Neural Information Processing Systems. 649657.Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. [49] Zhou Xiao, Mascolo Cecilia, and Zhao Zhongxiang. 2019. Topic-enhanced memory networks for personalised point-of-interest recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. ACM, 30183028.Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. [50] Zuo Yuan, Li Congrui, Lin Hao, and Wu Junjie. 2021. Topic modeling of short texts: A pseudo-document view with word embedding enhancement. IEEE Transactions on Knowledge and Data Engineering (2021), 11. Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Self-supervised Short-text Modeling through Auxiliary Context Generation

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in

          Full Access

          • Published in

            cover image ACM Transactions on Intelligent Systems and Technology
            ACM Transactions on Intelligent Systems and Technology  Volume 13, Issue 3
            June 2022
            415 pages
            ISSN:2157-6904
            EISSN:2157-6912
            DOI:10.1145/3508465
            • Editor:
            • Huan Liu
            Issue’s Table of Contents

            Copyright © 2022 Copyright held by the owner/author(s).

            This work is licensed under a Creative Commons Attribution International 4.0 License.

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 12 April 2022
            • Accepted: 1 January 2022
            • Revised: 1 December 2021
            • Received: 1 August 2021
            Published in tist Volume 13, Issue 3

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article
            • Refereed

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format .

          View HTML Format