|
{ |
|
"paper_id": "2022", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T16:22:13.468546Z" |
|
}, |
|
"title": "Generating Complement Data for Aspect Term Extraction with GPT-2", |
|
"authors": [ |
|
{ |
|
"first": "Amir", |
|
"middle": [], |
|
"last": "Pouran", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Ben", |
|
"middle": [], |
|
"last": "Veyseh", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Oregon", |
|
"location": { |
|
"settlement": "Eugene", |
|
"region": "Oregon", |
|
"country": "USA" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Franck", |
|
"middle": [], |
|
"last": "Dernoncourt", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Bonan", |
|
"middle": [], |
|
"last": "Min", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Thien", |
|
"middle": [ |
|
"Huu" |
|
], |
|
"last": "Nguyen", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Oregon", |
|
"location": { |
|
"settlement": "Eugene", |
|
"region": "Oregon", |
|
"country": "USA" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Aspect Term Extraction (ATE) is the task of identifying the word(s) in a review text toward which the author express an opinion. A major challenges for ATE involve data scarcity that hinder the training of deep sequence taggers to identify rare targets. To overcome these issues, we propose a novel method to better exploit the available labeled data for ATE by computing effective complement sentences to augment the input data and facilitate the aspect term prediction. In particular, we introduce a multistep training procedure that first obtains optimal complement representations and sentences for training data with respect to a deep ATE model. Afterward, we fine-tune the generative language model GPT-2 to allow complement sentence generation at test data. The REINFORCE algorithm is employed to incorporate different expected properties into the reward function to perform the fine-tuning. We perform extensive experiments on the benchmark datasets to demonstrate the benefits of the proposed method that achieve the state-of-the-art performance on different datasets.", |
|
"pdf_parse": { |
|
"paper_id": "2022", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Aspect Term Extraction (ATE) is the task of identifying the word(s) in a review text toward which the author express an opinion. A major challenges for ATE involve data scarcity that hinder the training of deep sequence taggers to identify rare targets. To overcome these issues, we propose a novel method to better exploit the available labeled data for ATE by computing effective complement sentences to augment the input data and facilitate the aspect term prediction. In particular, we introduce a multistep training procedure that first obtains optimal complement representations and sentences for training data with respect to a deep ATE model. Afterward, we fine-tune the generative language model GPT-2 to allow complement sentence generation at test data. The REINFORCE algorithm is employed to incorporate different expected properties into the reward function to perform the fine-tuning. We perform extensive experiments on the benchmark datasets to demonstrate the benefits of the proposed method that achieve the state-of-the-art performance on different datasets.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Aspect Term Extraction (ATE) is one of the fundamental tasks in Aspect-based Sentiment Analysis (ABSA). Its goal is to recognize the terms upon which a sentiment opinion is expressed in text. For instance, in the sentence \"The staff of the restaurant were good but the quality of the food was terrible\", an ATE system should recognize the two aspect terms (targets) \"staff \" and \"quality of food\". ATE finds its applications in ABSA systems to identify targets toward which sentiment analysis is done.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "A major challenge for ATE is the lack of enough training data. For instance, the widely-used Se-mEval datasets, e.g., Res15 (Pontiki et al., 2015) or Res16 (Pontiki et al., 2016) , contain less than 2,000 training samples with only 20% of the words appearing more than five times (Chen and Qian, 2020a ). This small size of training data hinders the deep sequence taggers to achieve optimal performance, especially for the tail targets (i.e., targets with few examples in the dataset) (He et al., 2018; Chen and Qian, 2019) . In order to alleviate this issue, prior work has resorted to data augmentation techniques to exploit additional training signals from different sources, including data from related tasks, e.g., ABSA (performing multi-tasking learning (Luo et al., 2020; Chen and Qian, 2020b) ), new labeled data for ATE produced by pre-trained sequence-to-sequence models , and soft prompts that are generated by pre-trained language models (Chen and Qian, 2020a) . As such, the critical requirements for such prior methods involve annotation for related tasks of ATE (e.g., ABSA), or large in-domain corpora to train the sequence-to-sequence/language models for data generation (called external data). Unfortunately, these requirements might be unavailable or very expensive to obtain in different domains, making it less applicable for various scenarios in practice.", |
|
"cite_spans": [ |
|
{ |
|
"start": 124, |
|
"end": 146, |
|
"text": "(Pontiki et al., 2015)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 156, |
|
"end": 178, |
|
"text": "(Pontiki et al., 2016)", |
|
"ref_id": "BIBREF23" |
|
}, |
|
{ |
|
"start": 280, |
|
"end": 301, |
|
"text": "(Chen and Qian, 2020a", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 485, |
|
"end": 502, |
|
"text": "(He et al., 2018;", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 503, |
|
"end": 523, |
|
"text": "Chen and Qian, 2019)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 760, |
|
"end": 778, |
|
"text": "(Luo et al., 2020;", |
|
"ref_id": "BIBREF16" |
|
}, |
|
{ |
|
"start": 779, |
|
"end": 800, |
|
"text": "Chen and Qian, 2020b)", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 950, |
|
"end": 972, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "To this end, this work aims to solve the issue of data scarcity for ATE without relying on annotated data for related tasks and large in-domain corpora. In particular, our main proposal is to fine-tune existing large-scale language models so they can generate complement sentences for input sentences in existing labeled datasets for ATE (i.e., not using external data as in prior works). Here, the motivation is that data scarcity might present a challenge for ATE models, especially on tail examples with rare aspect terms and context patterns (Chen and Qian, 2020a) . The complement sentence thus aims to provide supporting evidence and facilitate the recognition of aspect terms for the input sentences.", |
|
"cite_spans": [ |
|
{ |
|
"start": 546, |
|
"end": 568, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "As such, our method first seeks to obtain complement sentences for all the sentences in a given ATE dataset via a multi-step training procedure. In the first step, we train a base ATE model on a labeled training dataset to encode the available knowledge about aspect terms in the dataset. However, due to data scarcity, the base model might not be exposed sufficiently to aspect term patterns, thus limiting the ability of the produced representations for the input sentences to fully capture relevant information/features for ATE. To achieve complement sentences for ATE datasets, in the second step, we thus propose to learn optimal representation vectors/word embeddings that can be combined (e.g., via adding) to improve the representation vectors from the base model for ATE (called complement representations). Our motivation is that the insufficient coverage of aspect term patterns in the representations would cause the base model to exhibit poor performance (i.e., high loss) on the validation dataset. To this end, we propose to infer the complement representations for each validation sentence by incorporating them into the base model as additional parameters and minimizing the loss of the augmented model on the validation data. In the implementation, we divide the training data for an ATE dataset into k folds. By choosing one fold as validation data and treating the k \u2212 1 remaining folds as training data, the aforementioned procedure can return the optimal complement representations for each sentence in the validation fold. As such, we repeat this process for k possible choices of the validation fold that in all produce complement representations for each sentence in the training data.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "To employ the complement representations for training data, we can introduce them into the base model for retraining. However, this will cause a mismatch in the test time where labels for sentences are not available and the complement representations cannot be obtained. To this end, instead of directly using the learned complement representations, we propose to first transform them into complement sentences based on the GloVe word embeddings (Pennington et al., 2014) . This is done by introducing constraints to encourage the learned representation vectors to belong to the same space with GloVe word embeddings. The complement representations can then be mapped into complement sentences by finding the words whose GloVe embeddings are closest to the complement representations. In this way, each original training sentence for ATE can be associated with a complement sentence. Using pairs of original and complement sentences as training data, in the next step, we propose to train a generative model that can trans-form the original sentences into their complement versions. As such, in the test time, we can apply the generative model to generate complement sentences for test data and use GloVe embeddings to produce complement representation vectors for data augmentation.", |
|
"cite_spans": [ |
|
{ |
|
"start": 446, |
|
"end": 471, |
|
"text": "(Pennington et al., 2014)", |
|
"ref_id": "BIBREF22" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this work, we propose to fine-tune the language model GPT-2 (Radford et al., 2019) on the original and complement sentence pairs to obtain the generative model. Our motivation stems from the small number of the pairs for the original and complement sentences (due to the small size of ATE datasets) that might not be sufficient to train a generative model well. By leveraging the pretrained GPT-2 model, we expect that its language priors can compensate for the data scarcity issue and bootstrap the learning process from complement data. Finally, we use REINFORCE (Williams, 1992) to fine-tune GPT-2 to facilitate the enforcement of expected properties for the generated sentences (i.e, the similarity or the length comparability with respect to the complement sentences produced in prior step). We perform extensive evaluations for the proposed method on different benchmark datasets for ATE. Our experiments reveal the superior performance and demonstrate the effectiveness of the proposed method.", |
|
"cite_spans": [ |
|
{ |
|
"start": 63, |
|
"end": 85, |
|
"text": "(Radford et al., 2019)", |
|
"ref_id": "BIBREF31" |
|
}, |
|
{ |
|
"start": 568, |
|
"end": 584, |
|
"text": "(Williams, 1992)", |
|
"ref_id": "BIBREF39" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Problem Definition: ATE is formulated as a sequence labeling problem. Formally, given the input sentence S = [w 1 , w 2 , . . . , w n ], the goal is to predict the gold label sequence Y = [y 1 , y 2 , . . . , y n ] where y i \u2208 {B, I, O}, B stands for the \"Beginning of a target\", I stands for \"Inside a target\", and O stands for \"Other\". Our proposed model consists of a four-step procedure: (I) Training a base model for ATE using the available labeled data, (II) Finding the word representations of the optimal complement sentences for training data, (III) Fine-tuning a the language model GPT-2 to produce complement sentences for input sentences, and (IV) Training a final ATE model on the training data augmented with complement sentences.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "For the first step, we train a base model on an available labeled ATE dataset. The trained model will serve as a base to find the optimal complement representations for input sentences of the ATE dataset in the next step. To this end, we employ a Bi-LSTM base model. In particular, the input sentence S is first fed into the pre-trained BERT model (Devlin et al., 2019) to obtain the contextualized word embeddings X = [x 1 , x 2 , . . . , x n ] (x i is the average of the representation vectors for the wordpieces of w i in the last layer of BERT). As such, to further abstract the word embeddings X for ATE, we feed X into a Bidirectional LSTM (Bi-LSTM) network to obtain the hidden states", |
|
"cite_spans": [ |
|
{ |
|
"start": 348, |
|
"end": 369, |
|
"text": "(Devlin et al., 2019)", |
|
"ref_id": "BIBREF7" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Base ATE Model (Step I)", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "H = [h 1 , h 2 , . . . , h n ].", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Base ATE Model (Step I)", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Afterward, the vectors in H are sent into a twolayer feed-forward layer F F to generate the label probability distribution P (\u2022|S, w i ) for i-th word:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Base ATE Model (Step I)", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "P (\u2022|S, w i ) = F F (h i ).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Base ATE Model (Step I)", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Finally, to train the base model, we use negative log-likelihood loss:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Base ATE Model (Step I)", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "L b = \u2212 1 n n i=1 log P (y i |S, w i ).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Base ATE Model (Step I)", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "As mentioned in the introduction, the limited size of the ATE datasets might prevent the base model from being imposed sufficiently to training samples to learn necessary aspect term patterns in the representations, potentially leading to inferior performance (i.e., high loss on validation data), especially on tail targets. As such, it is necessary to enhance the representation learning capability of the base model by imposing it to further information. To achieve this goal, prior work has resorted to data augmentation or soft-prompts (Chen and Qian, 2020a) in which the training data is augmented with new sentences (e.g., generated by a pre-trained language model) to provide more evidences for aspect terms. However, the limitation of the prior work is that the generated sentences to augment ATE data is either ignorant of the ATE task (Chen and Qian, 2020a) or constrained on some heuristics (i.e., replacing non-aspect terms with other words generated by a language model) . As such, we argue that these data augmentation methods might not achieve the optimal augmentation for the available ATE data. We thus posit that the optimal augmentation for an input sentence is the one whose combination with the sentence could directly reduce the objective loss on validation data. Concretely, to find the optimal augmentation for a sentence S in the validation data, we search for the sentence S \u2032 whose combination with S (i.e., by adding their word representations) could further reduce the objective loss L b computed on validation data. Since this augmentation is optimized over validation data and not bound to any heuristics-based constraints, we expect it to be the optimal augmentation for the input sentence. Note that the optimality of the sentence S \u2032 is with respect to the objective loss L b and changing the criteria could lead to a different sentence S \u2032 .", |
|
"cite_spans": [ |
|
{ |
|
"start": 541, |
|
"end": 563, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 846, |
|
"end": 868, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "To find the optimal complement sentence S \u2032 for S in the validation data, since this is a discrete variable, we first attempt to find the representation vectors X \u2032 for its words w i . That is, S \u2032 is parameterized by a set of learnable vectors X \u2032 which are combined with the word embeddings X and are updated with the objective loss L b over validation data. In this work, the combination of X and X \u2032 is defended as the sum of their corresponding vectors x i and x \u2032", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "i . As such, the number of tokens of X \u2032 should be equal to the number of tokens of X, i.e.,", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "X \u2032 = [x \u2032 1 , x \u2032 2 , . . . , x \u2032 n ].", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "In addition, the dimension of the vectors should also match, i.e., |x", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "i | = |x \u2032 i | = D,", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "where D is the dimensionality of the word embedding vectors. Hence, the total number of parameters defined for all representation vectors is N \u00d7 n \u00d7 D, where N is the total number of sentences in the validation set.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "In the next step, we seek to optimize the representation parameters complement sentences by reducing the objective loss L b over validation data. In particular, for the sentence S with embeddings X and the complement sentence S \u2032 with parameters (i.e., embeddings) X \u2032 , we compute the sum of the corresponding vectors for the i-th token:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "x i = x i + \u03bbx \u2032", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "i where \u03bb is a trade-off parameter (i.e., the data augmentation in this work). The vectorsX = [x 1 ,x 2 , . . . ,x n ] will be sent to the base model architecture (i.e., BiLSTM followed by a feed-forward layer) to obtain the label distribution P (\u2022|S, S \u2032 , w i ). As such, the objective loss for this training step, i.e., L f , is defined similar to L b :", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "L f = \u2212 1 n n i=1 log P (y i |S, S \u2032 , w i ) (com- puted over validation data).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Note that in this training step, the original parameters of the trained base model is fixed so the only parameters to be updated are the parameters for the complement sentences S \u2032 , i.e., the vectors X \u2032 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Embedding Regularization: To further improve the complement embeddings and facilitate the mapping to complement sentences S \u2032 later, we introduce two additional regularization terms for the learning objective of complement embeddings. The first regularization seeks to encourage the complement embeddings X \u2032 to capture different (i.e., complementary) information from those for the embeddings X of the input sentence S, thus enhancing the contribution of complement embeddings. In particular, we compute the representation vectors R S and R S \u2032 for the original and complement sentences using the max-pooling operation:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "R S = M AX_P OOL(x 1 , x 2 , . . . , x n ) and R S \u2032 = M AX_P OOL(x \u2032 1 , x \u2032 2 , . . . , x \u2032 n )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": ". Afterward, the complementary nature of embeddings is enforced by introducing the dot product L reg between R S and R S \u2032 into the loss function for minimization (i.e., minimizing the similarity between R S and R S \u2032 ):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "L reg = R S \u2299 R S \u2032 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "For the second regularization, we aim to align the complement embeddings X \u2032 to the space of the GloVe embeddings (Pennington et al., 2014) to facilitate the transformation to complement sentences in the next step. Here, we use the GloVe embeddings for convenience and leave other possible pre-trained embeddings for future work. In particular, for each vector x \u2032 i \u2208 X \u2032 , we use a feedforward network F to transform x \u2032 i into the vector F (x \u2032 i ) of the same dimension with Glove embeddings. Afterward, we find the vector e i in the GloVe embedding table that is closest to F (x \u2032 i ) based on the Euclidean distance. The Euclidean distance between F (x \u2032 i ) and e i is then incorporated into the loss function to promote the alignment of complement and GloVe embeddings:", |
|
"cite_spans": [ |
|
{ |
|
"start": 114, |
|
"end": 139, |
|
"text": "(Pennington et al., 2014)", |
|
"ref_id": "BIBREF22" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "L GloV e = 1 n n i=1 \u2225F (x \u2032 i ) \u2212 e i \u2225 2 2 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Finally, the overall loss function to learn the complement representations X \u2032 is:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "L emb = L f + \u03b1 reg L reg + \u03b1 GloV e L GloV e", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "where \u03b1 reg and \u03b1 GloV e are the trade-off parameters. Note that the parameters for F are also optimized in this process.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "As such, this training step produces the complement embedding X \u2032 for each sentence in the validation data. To maximize the use of data, we implement this training step in a 10-fold validation fashion described in the introduction. In particular, we train the base model on 9 folds of the training data (i.e., Section 2.1) and use the remaining fold for the validation data in the complement representation optimization. By alternating the choice of validation fold, we can obtain a complement representation sequence X \u2032 for each sentence in the original training data.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Finding Complement Representations (Step II)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "As mentioned in the introduction, the complement embeddings X \u2032 can be used directly to augment training data and train a model for ATE. However, as the optimization for complement embeddings cannot be done in the test time (due to the unavailability of labels for data), the direct augmentation will cause a mismatch between the training and test phases. To enable the generation of complement embeddings in the test time, we thus propose to first transform the complement embeddings X \u2032 into a complement sentence", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "S \u2032 = [w \u2032 1 , w \u2032 2 , . . . , w \u2032 n ] where w \u2032", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "i is the word whose Glove embedding is closest to the transformed complement vector F (x \u2032 i ) for w i . The set of every pair (S, S \u2032 ) for sentences S in training data is then employed to train a generative language model that seeks to consume S and produce its complement sentence S \u2032 . In this way, we can apply the generative model in the test time to generate complement sentences for test data, that, in turn, can be transformed into complement embeddings by mapping words into Glove embedding vectors for data augmentation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "One potential issue is that the number of original and complement sentence pairs (S, S \u2032 ) might be small due to the limited size of ATE datasets, thus hindering the training of effective generative models for our complement sentence goal. As such, we propose to leverage the language priors in the pre-trained generative model GPT-2 (Radford et al., 2019) as the bootstrap knowledge for the complement generation. In particular, we propose to fine-tune the GPT-2 model on the sentence pairs (S, S \u2032 ) in this step. The policy-gradient method REINFORCE (Williams, 1992) is utilized for the fine-tuning process to facilitate the incorporation of different expected properties for complement sentences. Concretely, the input to GPT-2 consists of the S sentence \"w 1 w 2 . . . w n SEP \" from which GPT-2 will generate the sentence S \u2032\u2032 . To compute the reward for the generated sentence S \u2032\u2032 , we propose three objectives: (1) Similarity with Complement Sentence: The generated sentence S \u2032\u2032 should be similar to the actual complement sentence S \u2032 . To compute the similarity of the two sentences, we employ the CIDEr score (Vedantam et al., 2015) for S \u2032\u2032 : R sim = CIDEr(S \u2032\u2032 ); (2) Length Penalty: As discussed earlier, since we use sum of the corresponding word embeddings of the original and complement sentences for data augmentation, it is intuitive to encourage the generated sentences S \u2032\u2032 to have the same length as the original sentence S. Thus, we introduce the length penalty as a part of the reward: R len = ||S|\u2212|S \u2032\u2032 ||;", |
|
"cite_spans": [ |
|
{ |
|
"start": 334, |
|
"end": 356, |
|
"text": "(Radford et al., 2019)", |
|
"ref_id": "BIBREF31" |
|
}, |
|
{ |
|
"start": 553, |
|
"end": 569, |
|
"text": "(Williams, 1992)", |
|
"ref_id": "BIBREF39" |
|
}, |
|
{ |
|
"start": 1121, |
|
"end": 1144, |
|
"text": "(Vedantam et al., 2015)", |
|
"ref_id": "BIBREF36" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "(3) Difference with Original Sentence: Similar to embedding regularization L reg presented earlier for complement embeddings, here we also aim to promote the semantic difference between the generated sentence S \u2032\u2032 and the original sentence S (for complementary information). To this end, we represent each sentence using the max-pooled representation of their word embeddings obtained from the GloVe embedding table, i.e.,R S andR S \u2032\u2032 . Next, their dot-product is employed for the difference reward R dif f =R S \u2299R S \u2032\u2032 . Consequently, the overall reward to train the generative model is computes as R(S \u2032\u2032 ) = R sim \u2212 \u03b2R len \u2212 \u03b3R dif f . With REINFORCE, we seek to minimize the negative expected reward R(S \u2032\u2032 ) over the possible choices of S \u2032\u2032 :", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "L tune = \u2212E\u015c \u2032\u2032 \u223cP (\u015c \u2032\u2032 |S) [R(\u015c \u2032\u2032 )].", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "The policy gradient is then estimated by:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "\u2207L tune = \u2212E\u015c \u2032\u2032 \u223cP (\u015c \u2032\u2032 |S) [(R(\u015c \u2032\u2032 ) \u2212 b)\u2207 log P (\u015c \u2032\u2032 |S)]", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": ". Using one roll-out sample, we further estimate \u2207L tune via the generated sentence S \u2032\u2032 :", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "\u2207L tune = \u2212(R(S \u2032\u2032 ) \u2212 b)\u2207 log P (\u015c \u2032\u2032 |S)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "where b is the baseline to reduce variance. In this work, we obtain the baseline b via:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "b = 1 |B| |B| i=1 R(S \u2032\u2032 i ),", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "where |B| is the mini-batch size and S \u2032\u2032 i is the generated sentence for the i-th sample in the mini-batch.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Generating Complement Sentences (Step III)", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "To achieve a consistency in the training and testing phase, we use the generated sentences from the finetuned GPT-2 model as the complement sentences for data augmentation in both phases. In particular, for the training data, similar to the complement embedding optimization Section 2.2, the fine-tuning of GPT-2 is performed with 10-fold cross validation. In particular, the GPT-2 model is fine-tuned on the (S, S \u2032 ) pairs of 9 folds and then employed to generate S \u2032\u2032 for each sentence in the remaining fold. To this end, each sentence S in the training data is associated with a generated sentence S \u2032\u2032 . For test data, we simply apply the fine-tuned GPT-2 model directly to generate a complement sentence for each sentence in that data.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "As such, for each sentence S (in the training or test data), its complement sentence from GPT-2 ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "S \u2032\u2032 = [w \u2032\u2032 1 , w \u2032\u2032 2 , . . . , w \u2032\u2032 n ]", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "augmented representationX = [x 1 ,x 2 , . . . ,x n ]", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "from the two sentences is computed by the sum of their corresponding word representations:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "x i = x i + \u03bbG(x \u2032\u2032 i ),", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "where G is a feed-forward network to match the dimensions of the GloVe embedding G(x \u2032\u2032 i ) and x i . Finally, following the base ATE architecture, the resulting vectorsX are sent to a BiLSTM network followed by a feed-forward layer to obtain the label distribution P (\u2022|S, S \u2032\u2032 , w i ) for the i-th word. This distribution is used for prediction in the test phase while the training phase employs the negative log-likelihood (over training data) to train the final ATE model:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "L f inal = \u2212 1 n n i=1 log P (y i |S, S \u2032\u2032 , w i ).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Training a Final ATE Model (Step IV)", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "Datasets & Parameters: To evaluate the effectiveness of the proposed method, we employ the commonly used SemEval datasets for ATE. Specifically, we use the datasets of SemEval 2014 Task 4 (Pontiki et al., 2014) , i.e., Lap14 and Res14 reviews for laptops and restaurants, and SemEval 2015 Task 12 (Pontiki et al., 2015) and SemEval 2016 Task 5 (Pontiki et al., 2016) , i.e., Res15 and Res16 with reviews for restaurants. Following prior work (Chen and Qian, 2020a), we employ the official train/test splits and randomly select 150 samples from training data as validation data for those datasets. Table 1 shows the statistics for the datasets.", |
|
"cite_spans": [ |
|
{ |
|
"start": 188, |
|
"end": 210, |
|
"text": "(Pontiki et al., 2014)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 297, |
|
"end": 319, |
|
"text": "(Pontiki et al., 2015)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 344, |
|
"end": 366, |
|
"text": "(Pontiki et al., 2016)", |
|
"ref_id": "BIBREF23" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 597, |
|
"end": 605, |
|
"text": "Table 1", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experiments", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In our experiments, we use the development set of the Lap14 dataset to tune the hyper-parameters. Based on our experiments, the following values are selected: 200 dimensions for the BiLSTM layer and the feed-forward layers; 1 layer for BiLSTM and 2 layers for the feed-forward networks; the BERT base version for the encoding layer with fixed parameters; GPT-2 small for sentence generation; 0.3 for the \u03bb in the word vector augmentation; 0.1 for the trade-off parameters \u03b1 reg and \u03b1 GloV e in the complement embedding optimization; 0.1 and 0.05 for the trade-off parameters \u03b2 and \u03b3 in the reward for GPT-2 fine-tuning; 0.3 for the learning rate for the Adam optimizer; and 50 for batch-size. Baselines: Following prior work, we compare our model with: (1) the winners of the SemEval tasks: IHS-RD (Chernyshevich, 2014) , DLIREC (Toh and Wang, 2014) , EliXa (San Vicente et al., 2015), NLANGP (Toh and Su, 2016) ; (2) Deep joint models, i.e., jointly train ATE with Opinion Term Extraction (OTE) or Aspect-Based Sentiment Analysis (ABSA): RNCRF (Wang et al., 2016) , MIN (Li and Lam, 2017) , CMLA (Mao et al., 2021) , HAST (Li et al., 2018) , RACL (Chen and Qian, 2020b) , Dual-MRC (Mao et al., 2021) ; (3) Deep models trained on ATE datasets augmented with external in-domain corpora and resources: BiLSTM-CRF , Seq2Seq (Ma et al., 2019) , BERT , DE-CNN (Xu et al., 2018) , BERT-PT , SoftProto (Chen and Qian, 2020a) . We also compare with CL-BERT (Yang et al., 2020) which employs constituency trees for ATE. For the evaluation metric, following prior work, we report the F1 score for aspect term prediction. A prediction is counted as correct if its boundaries match the gold aspect term. We name our model ATEOA which stands for Aspect Term Extraction with Optimal Augmentation.", |
|
"cite_spans": [ |
|
{ |
|
"start": 798, |
|
"end": 819, |
|
"text": "(Chernyshevich, 2014)", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 829, |
|
"end": 849, |
|
"text": "(Toh and Wang, 2014)", |
|
"ref_id": "BIBREF35" |
|
}, |
|
{ |
|
"start": 893, |
|
"end": 911, |
|
"text": "(Toh and Su, 2016)", |
|
"ref_id": "BIBREF34" |
|
}, |
|
{ |
|
"start": 1045, |
|
"end": 1064, |
|
"text": "(Wang et al., 2016)", |
|
"ref_id": "BIBREF38" |
|
}, |
|
{ |
|
"start": 1071, |
|
"end": 1089, |
|
"text": "(Li and Lam, 2017)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 1097, |
|
"end": 1115, |
|
"text": "(Mao et al., 2021)", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 1123, |
|
"end": 1140, |
|
"text": "(Li et al., 2018)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 1148, |
|
"end": 1170, |
|
"text": "(Chen and Qian, 2020b)", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 1182, |
|
"end": 1200, |
|
"text": "(Mao et al., 2021)", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 1321, |
|
"end": 1338, |
|
"text": "(Ma et al., 2019)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 1355, |
|
"end": 1372, |
|
"text": "(Xu et al., 2018)", |
|
"ref_id": "BIBREF41" |
|
}, |
|
{ |
|
"start": 1395, |
|
"end": 1417, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 1449, |
|
"end": 1468, |
|
"text": "(Yang et al., 2020)", |
|
"ref_id": "BIBREF42" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Results: The main results are shown in Table 2 . This table shows that the proposed model can effectively improve the performance compared to existing joint inference and data augmentation methods. This achievement is significant as the proposed model does not utilize any external in-domain data nor extra supervision from other related tasks. This is important for domains with limited data where collecting large-scale in-domain data or supervision from related tasks could be prohibitively expensive. Moreover, as the proposed model employs pre-trained language models (i.e., GPT-2) to generate effective augmentation sentences for training/test data, it can directly benefit from growing advances in pre-trained language models.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 39, |
|
"end": 46, |
|
"text": "Table 2", |
|
"ref_id": "TABREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experiments", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Ablation Study: The proposed ATEOA model is trained in four major training steps. In this section, we study the role of those proposed steps for the performance of the ATE model. For each training step, we aim to answer two questions: (i) Whether the proposed step is beneficial for ATEOA? and (ii) Is the current configuration for the step optimal? To this end, we consider the following ablated models:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "(1) -Base Model Training (Step I): This model ignores step I to train a base ATE model in Section 2.1. In particular, step II for finding complement embeddings will only employ an ATE base model with randomly initialized parameters. Here, the parameters for the base model are not fixed; they are jointly updated with the complement embeddings in step II of the training; (2) -Complement Representation Finding (Step II): This model excludes step II of the training procedure that makes the optimized complement representations unavailable for the fine-tuning of GPT-2 in step III. As such, to achieve a fair access to the trained base model in step I in this model, we change step III by fine-tuning the GPT-2 model with the reward of F1 score of the trained base model (from step I) on the validation data. Here, the base model is applied on the representation combinations of the original and GPT-generated sentences (also using GloVe embeddings for the words in the generated sentences); (3) -Embedding Regularization L reg : This model removes the regularization loss L reg in step II of the training process; (4) -GloVe Alignment L GloV e : This model excludes the regularization L GloV e for representation alignment with GloVe in step II; (5) -Language Model (Step III): This ablated models does not utilize step III, thus eliminating the GPT-2 model trained over the original and complement sentence pairs (S, S \u2032 ). As such, in step IV, we directly retrain the base model on the augmented data with the complement representations X \u2032 (i.e., x i + \u03bbx \u2032 i ) and do not apply data augmentation for test data (i.e., applying the train model on x i directly); (6) -Language Model + FForward: This model is similar to (5) (i.e., excluding GPT-2 in step III). However, to allow the augmentation on test data, a feed-forward network is trained on pairs (x i , x \u2032 i ) to directly transform the representation vectors x i of the original sentence S into the complement representations for data augmentation in both training and test phases of step IV; (7) -Similarity Reward: For this model, we do not use the similarity reward R sim in the reward function to fine-tune GPT-2 in step III; (8) -Length Penalty: This model does not employ the length penalty R len in the reward for tuning GPT-2; (9) -Difference Reward: For this baseline, the reward based on difference with original sentence, i.e., R dif f , is ablated from the reward for GPT-2 fine-tuning; (10) -Final Training (Step IV): This baseline skips the last step of the proposed training procedure. As such, the combined representations of the original sentence S and the complement sentence S \u2032\u2032 generated by GPT-2 (i.e., x i + \u03bbG(x \u2032\u2032 i )) are directly sent into the base ATE model (trained over the entire training data) from step I for prediction; and (11) -Generated Data in Final Training: Finally, to demonstrate the benefit of augmenting training data with generated sentences from the fine-tuned GPT-2 model in step IV, we report the performance of the base model that is instead trained on the combination of the word representations X and the optimized complement representations X \u2032 , i.e., x i + \u03bbx \u2032 i in step II (as in (5)). The fine-tuned GPT-2 model is still used to generate complement sentences for data augmentation in the test phase for this model. The performance of the models on the test sets of the SemEval datasets is reported in Table 3 . This table clearly shows that all training steps in the proposed procedure are necessary as skipping any of these steps will hurt the performance. In particular, among the four steps, removing step III has the most negative impact as the ablated model \"-Language Model\" has the lowest performance across datasets. We attribute the importance of step III to its ability to enable augmentation consistency for training and test data (i.e., the fine-tuned GPT-2 can generate complement sentences for both training and test data). This is further highlighted by the worse performance of the \"-Generated Data in Final Training\" model where the training data is augmented with X \u2032 , but GPT-generated data is used to augment test data. Table 3 also shows that among the three awards for GPT-2 fine-tuning, the similarity reward is most important. This is expected as the primary goal of fine-tuning is to gen- ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 3418, |
|
"end": 3425, |
|
"text": "Table 3", |
|
"ref_id": "TABREF5" |
|
}, |
|
{ |
|
"start": 4158, |
|
"end": 4165, |
|
"text": "Table 3", |
|
"ref_id": "TABREF5" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experiments", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Generative Language Models: As it is evident in the ablation study, exploiting a pre-trained generative model (i.e., GPT-2) for ATEOA is preferable since it can provide language priors to support the sentence generation learning from limited ATE datasets. In this section, we study how the performance of the model changes if we alter the generative language model used in step III of ATEOA. Concretely, we compare the performance of three different models: (1) GPT-2 (Radford et al., 2019): This transformer-based model is pre-trained on WebText corpus. We examine its small version with 117 million parameters;", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Analysis", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "(2) T5 (Raffel et al., 2019 ): This language model employs the encoder-decoder architecture in Transformer for sequence-to-sequence tasks. We explore its base version with 220 million parameters. We use the input sentence S as the source sequence and the complement sentence S \u2032 as the target sequence to fine-tune the T5 model; and (3) BART (Lewis et al., 2019) : This model is a transformer-based auto-encoder language model. We also utilize its base version with 139 million parameters. Similar to T5, this is a sequence-to-sequence generative model that is fine-tuned by treating S and S \u2032 as the source and target sequences respectively. To compare the performance of the three language models, we use them in the training step III of the final ATE model and report the corresponding performance. Furthermore, we compare the language models on their capability to generate sentences that are similar to the complement sentences S \u2032 . In particular, using the Lap14 dataset, we seek to find a complement sentence S \u2032 for each sentence in the test data portion with the proposed method. Table 5 : Similarity between the generated sentences from the language models and the \"ground-truth\" complement sentences on test data of Lap14.", |
|
"cite_spans": [ |
|
{ |
|
"start": 7, |
|
"end": 27, |
|
"text": "(Raffel et al., 2019", |
|
"ref_id": "BIBREF32" |
|
}, |
|
{ |
|
"start": 342, |
|
"end": 362, |
|
"text": "(Lewis et al., 2019)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 1090, |
|
"end": 1097, |
|
"text": "Table 5", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Analysis", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "To this end, a base model is first trained on the training data portion using step I; the complement representations X \u2032 are then computed for each sentence in the test data portion using step II; each X \u2032 is then mapped into the complement sentence S \u2032 with the GloVe embeddings. Here, S \u2032 serves as the \"ground-truth\" complement sentence for the test sentences in our approach. Next, we use the finetuned language model to generate the complement sentence S \u2032\u2032 for each test sentence (i.e., prompting the language model with test data). Finally, we evaluate the similarity of the generated sentence S \u2032\u2032 and the \"ground-truth\" complement sentence S \u2032 (for the test data) using ROUGE-1, ROUGE-2, METEOR (Banerjee and Lavie, 2005) and BLUE4 as the similarity metrics. The results for this experiment are shown in Tables 4 and 5 . Both tables clearly demonstrate the capacity of GPT-2 to generate better complement sentences to augment ATE data (i.e., yielding better performance for ATE in Table 4 and generating more similar sentences to the obtained complement sentences S \u2032 in Table 5 ).", |
|
"cite_spans": [ |
|
{ |
|
"start": 704, |
|
"end": 730, |
|
"text": "(Banerjee and Lavie, 2005)", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 813, |
|
"end": 827, |
|
"text": "Tables 4 and 5", |
|
"ref_id": "TABREF7" |
|
}, |
|
{ |
|
"start": 990, |
|
"end": 997, |
|
"text": "Table 4", |
|
"ref_id": "TABREF7" |
|
}, |
|
{ |
|
"start": 1080, |
|
"end": 1087, |
|
"text": "Table 5", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Analysis", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Tail Aspect Term Analysis: Following prior work (Chen and Qian, 2020a) , we evaluate the performance for our model on the tail aspect terms in test data (i.e., aspect terms occurring less than 5 times in the training sets). As such, we compare our model with prior work that reports their performance in this analysis, i.e., DE-CNN (Xu et al., 2018) and SoftProto (Chen and Qian, 2020a) . Note that we replace the contextualized BERT representations (i.e., X) in our model with the GloVe embeddings to achieve a fair comparison with prior work in this section. The results are provided in Table 6 that clearly shows the superiority of ATEOA to recognize tail aspect terms and further highlights the benefits of the proposed method. Case Study: To provide more insight into the quality of the complement sentences generated by the pre-trained GPT-2 model, Table 7 showcases some examples from the laptop and restaurant domains whose aspect terms can only be correctly predicted by our proposed method (i.e., prior work fails to recognize aspect terms in these cases). The table suggests that although the generated sentences might not look natural, they clearly provide more evidence and emphasis on the aspect terms which makes the task easier for the ATE model. Specifically, in the first example, the generated complement sentence emphasizes the target word \"touchpad\" in the original sentence by replicating it and including the related word \"mouse\". The same pattern of emphasis can be seen in the second example where the model excludes the word \"money\" and includes the related word \"Food\" (that are more related to the target word \"meal\") in the generated sentence.", |
|
"cite_spans": [ |
|
{ |
|
"start": 48, |
|
"end": 70, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 332, |
|
"end": 349, |
|
"text": "(Xu et al., 2018)", |
|
"ref_id": "BIBREF41" |
|
}, |
|
{ |
|
"start": 364, |
|
"end": 386, |
|
"text": "(Chen and Qian, 2020a)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 855, |
|
"end": 862, |
|
"text": "Table 7", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Analysis", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "ATE has been first approached with rule-based (Hu and Liu, 2004; Wu et al., 2009) or featurebased (Li et al., 2010; Chen et al., 2014; Toh and Su, 2016) methods. Recently, ATE methods have focused on neural networks such as LSTM (Liu et al., 2015) , CNN (Xu et al., 2018) or Transformer ). An ATE system can be used in downstream applications such as sentiment analysis Pouran Ben Veyseh et al., 2020b; Orbach et al., 2021) or opinion term extraction (Pouran Ben Veyseh et al., 2020a) . One of the challenges for this task is the scarcity of training data which hinders the training of large neural networks. To alleviate this issue, two major directions have been explored in the literature: (I) Joint Training, i.e., jointly solving ATE task with another related task such as ABSA (Wang et al., 2016; Mao et al., 2021; Chen and Qian, 2020b) or Opinion Term Extraction (OTE) (Li and Lam, 2017; Dai and Song, 2019) ; and (II) Data Augmentation, i.e., augmenting ATE models with in-domain unlabeled data (Xu et al., 2018; Chen and Qian, Input Sentences Generated Complement Sentences Targets however , there are major issues with Although , exist some problems with the touchpad touchpad the touchpad which render the device nearly useless . and mouse makes touchpad useless and touchpad useless way too much money for such a terrible meal .", |
|
"cite_spans": [ |
|
{ |
|
"start": 46, |
|
"end": 64, |
|
"text": "(Hu and Liu, 2004;", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 65, |
|
"end": 81, |
|
"text": "Wu et al., 2009)", |
|
"ref_id": "BIBREF40" |
|
}, |
|
{ |
|
"start": 98, |
|
"end": 115, |
|
"text": "(Li et al., 2010;", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 116, |
|
"end": 134, |
|
"text": "Chen et al., 2014;", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 135, |
|
"end": 152, |
|
"text": "Toh and Su, 2016)", |
|
"ref_id": "BIBREF34" |
|
}, |
|
{ |
|
"start": 229, |
|
"end": 247, |
|
"text": "(Liu et al., 2015)", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 254, |
|
"end": 271, |
|
"text": "(Xu et al., 2018)", |
|
"ref_id": "BIBREF41" |
|
}, |
|
{ |
|
"start": 370, |
|
"end": 402, |
|
"text": "Pouran Ben Veyseh et al., 2020b;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 403, |
|
"end": 423, |
|
"text": "Orbach et al., 2021)", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 451, |
|
"end": 484, |
|
"text": "(Pouran Ben Veyseh et al., 2020a)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 783, |
|
"end": 802, |
|
"text": "(Wang et al., 2016;", |
|
"ref_id": "BIBREF38" |
|
}, |
|
{ |
|
"start": 803, |
|
"end": 820, |
|
"text": "Mao et al., 2021;", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 821, |
|
"end": 842, |
|
"text": "Chen and Qian, 2020b)", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 876, |
|
"end": 894, |
|
"text": "(Li and Lam, 2017;", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 895, |
|
"end": 914, |
|
"text": "Dai and Song, 2019)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 1003, |
|
"end": 1020, |
|
"text": "(Xu et al., 2018;", |
|
"ref_id": "BIBREF41" |
|
}, |
|
{ |
|
"start": 1021, |
|
"end": 1051, |
|
"text": "Chen and Qian, Input Sentences", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Food costs so much for such a bad meal . meal Table 7 : Generated complement sentences by 2020a). In this work, we also propose a method to augment the training data for ATE. However, unlike prior work that requires large in-domain corpus, our approach employs an existing largescale language model (i.e., to facilitate the generation of complement sentences for the available ATE datasets. Using GPT-2 to address data scarcity has been shown to be effective in other domains and tasks (Papanikolaou and Pierleoni, 2020; Pouran Ben Veyseh et al., 2021; Peng et al., 2020) . In this work, we demonstrate the viability of this technique for aspect term extraction.", |
|
"cite_spans": [ |
|
{ |
|
"start": 486, |
|
"end": 520, |
|
"text": "(Papanikolaou and Pierleoni, 2020;", |
|
"ref_id": "BIBREF20" |
|
}, |
|
{ |
|
"start": 521, |
|
"end": 552, |
|
"text": "Pouran Ben Veyseh et al., 2021;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 553, |
|
"end": 571, |
|
"text": "Peng et al., 2020)", |
|
"ref_id": "BIBREF21" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 46, |
|
"end": 53, |
|
"text": "Table 7", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "We introduce a new training procedure for Aspect Term Extraction. In the proposed procedure, the available ATE dataset is employed to train a deep model which is further used to find complement representations for input sentences in training data. Later, to obtain the complement sentences at the inference time, we tine-tune the pre-trained language model GPT-2 to generate sentences similar to the complement sentences found in the previous steps (with GloVe mapping). Our extensive experiments on benchmark datasets reveal the superiority of the proposed model, leading to the state-of-the-art performance for the datasets. Moreover, our analysis show that all steps of the proposed procedure are necessary and effective for the ATE task. In future, we will explore the application of this procedure in other related task such as OTE and ABSA.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "6" |
|
}, |
|
{ |
|
"text": "Note that the generated sentence S \u2032\u2032 might have a different length from the original sentence S. For these cases, we pad (with zero vectors) or truncate the vector sequence X \u2032\u2032 to have the same length as S.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "This research has been supported by the Army Research Office (ARO) grant W911NF-21-1-0112 and the NSF grant CNS-1747798 to the IU-CRC Center for Big Learning. This research is also based upon work supported by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA), via IARPA Contract No. 2019-19051600006 under the Better Extraction from Text Towards Enhanced Retrieval (BETTER) Program. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies, either expressed or implied, of ARO, ODNI, IARPA, the Department of Defense, or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright annotation therein. This document does not contain technology or technical data controlled under either the U.S. International Traffic in Arms Regulations or the U.S. Export Administration Regulations.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgement", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Meteor: An automatic metric for mt evaluation with improved correlation with human judgments", |
|
"authors": [ |
|
{ |
|
"first": "Satanjeev", |
|
"middle": [], |
|
"last": "Banerjee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alon", |
|
"middle": [], |
|
"last": "Lavie", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "65--72", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Satanjeev Banerjee and Alon Lavie. 2005. Meteor: An automatic metric for mt evaluation with improved cor- relation with human judgments. In Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summariza- tion, pages 65-72.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Aspect extraction with automated prior knowledge learning", |
|
"authors": [ |
|
{ |
|
"first": "Zhiyuan", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Arjun", |
|
"middle": [], |
|
"last": "Mukherjee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bing", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "347--358", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.3115/v1/P14-1033" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhiyuan Chen, Arjun Mukherjee, and Bing Liu. 2014. Aspect extraction with automated prior knowledge learning. In Proceedings of the 52nd Annual Meet- ing of the Association for Computational Linguistics (Volume 1: Long Papers), pages 347-358, Baltimore, Maryland. Association for Computational Linguis- tics.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Transfer capsule network for aspect level sentiment classification", |
|
"authors": [ |
|
{ |
|
"first": "Zhuang", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tieyun", |
|
"middle": [], |
|
"last": "Qian", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "547--556", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/P19-1052" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhuang Chen and Tieyun Qian. 2019. Transfer capsule network for aspect level sentiment classification. In Proceedings of the 57th Annual Meeting of the As- sociation for Computational Linguistics, pages 547- 556, Florence, Italy. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Enhancing aspect term extraction with soft prototypes", |
|
"authors": [ |
|
{ |
|
"first": "Zhuang", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tieyun", |
|
"middle": [], |
|
"last": "Qian", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "2107--2117", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.emnlp-main.164" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhuang Chen and Tieyun Qian. 2020a. Enhancing as- pect term extraction with soft prototypes. In Proceed- ings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2107-2117, Online. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Relation-aware collaborative learning for unified aspect-based sentiment analysis", |
|
"authors": [ |
|
{ |
|
"first": "Zhuang", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tieyun", |
|
"middle": [], |
|
"last": "Qian", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "3685--3694", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.acl-main.340" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhuang Chen and Tieyun Qian. 2020b. Relation-aware collaborative learning for unified aspect-based sen- timent analysis. In Proceedings of the 58th Annual Meeting of the Association for Computational Lin- guistics, pages 3685-3694, Online. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "IHS R&D Belarus: Cross-domain extraction of product features using CRF", |
|
"authors": [ |
|
{ |
|
"first": "Maryna", |
|
"middle": [], |
|
"last": "Chernyshevich", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 8th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "309--313", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.3115/v1/S14-2051" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Maryna Chernyshevich. 2014. IHS R&D Belarus: Cross-domain extraction of product features using CRF. In Proceedings of the 8th International Work- shop on Semantic Evaluation (SemEval 2014), pages 309-313, Dublin, Ireland. Association for Computa- tional Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Neural aspect and opinion term extraction with mined rules as weak supervision", |
|
"authors": [ |
|
{ |
|
"first": "Hongliang", |
|
"middle": [], |
|
"last": "Dai", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yangqiu", |
|
"middle": [], |
|
"last": "Song", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "5268--5277", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/P19-1520" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hongliang Dai and Yangqiu Song. 2019. Neural aspect and opinion term extraction with mined rules as weak supervision. In Proceedings of the 57th Annual Meet- ing of the Association for Computational Linguistics, pages 5268-5277, Florence, Italy. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
|
"authors": [ |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Devlin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming-Wei", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kenton", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kristina", |
|
"middle": [], |
|
"last": "Toutanova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "4171--4186", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/N19-1423" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Tech- nologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, Minnesota. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Exploiting document knowledge for aspect-level sentiment classification", |
|
"authors": [ |
|
{ |
|
"first": "Ruidan", |
|
"middle": [], |
|
"last": "He", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hwee Tou", |
|
"middle": [], |
|
"last": "Wee Sun Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Ng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Dahlmeier", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "579--585", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/P18-2092" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ruidan He, Wee Sun Lee, Hwee Tou Ng, and Daniel Dahlmeier. 2018. Exploiting document knowledge for aspect-level sentiment classification. In Proceed- ings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 579-585, Melbourne, Australia. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Mining and summarizing customer reviews", |
|
"authors": [ |
|
{ |
|
"first": "Minqing", |
|
"middle": [], |
|
"last": "Hu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bing", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "168--177", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Minqing Hu and Bing Liu. 2004. Mining and summa- rizing customer reviews. In Proceedings of the tenth ACM SIGKDD international conference on Knowl- edge discovery and data mining, pages 168-177.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension", |
|
"authors": [ |
|
{ |
|
"first": "Mike", |
|
"middle": [], |
|
"last": "Lewis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yinhan", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Naman", |
|
"middle": [], |
|
"last": "Goyal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Marjan", |
|
"middle": [], |
|
"last": "Ghazvininejad", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Abdelrahman", |
|
"middle": [], |
|
"last": "Mohamed", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Omer", |
|
"middle": [], |
|
"last": "Levy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ves", |
|
"middle": [], |
|
"last": "Stoyanov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luke", |
|
"middle": [], |
|
"last": "Zettlemoyer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"arXiv": [ |
|
"arXiv:1910.13461" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. Bart: De- noising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Structure-aware review mining and summarization", |
|
"authors": [ |
|
{ |
|
"first": "Fangtao", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chao", |
|
"middle": [], |
|
"last": "Han", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Minlie", |
|
"middle": [], |
|
"last": "Huang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaoyan", |
|
"middle": [], |
|
"last": "Zhu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ying-Ju", |
|
"middle": [], |
|
"last": "Xia", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shu", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hao", |
|
"middle": [], |
|
"last": "Yu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 23rd International Conference on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "653--661", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fangtao Li, Chao Han, Minlie Huang, Xiaoyan Zhu, Ying-Ju Xia, Shu Zhang, and Hao Yu. 2010. Structure-aware review mining and summarization. In Proceedings of the 23rd International Conference on Computational Linguistics (Coling 2010), pages 653-661, Beijing, China. Coling 2010 Organizing Committee.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Conditional augmentation for aspect term extraction via masked sequence-tosequence generation", |
|
"authors": [ |
|
{ |
|
"first": "Kun", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chengbo", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaojun", |
|
"middle": [], |
|
"last": "Quan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Qing", |
|
"middle": [], |
|
"last": "Ling", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yan", |
|
"middle": [], |
|
"last": "Song", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "7056--7066", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.acl-main.631" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kun Li, Chengbo Chen, Xiaojun Quan, Qing Ling, and Yan Song. 2020. Conditional augmentation for aspect term extraction via masked sequence-to- sequence generation. In Proceedings of the 58th An- nual Meeting of the Association for Computational Linguistics, pages 7056-7066, Online. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Aspect term extraction with history attention and selective transformation", |
|
"authors": [ |
|
{ |
|
"first": "Xin", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lidong", |
|
"middle": [], |
|
"last": "Bing", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Piji", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wai", |
|
"middle": [], |
|
"last": "Lam", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Zhimou", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "IJCAI", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xin Li, Lidong Bing, Piji Li, Wai Lam, and Zhimou Yang. 2018. Aspect term extraction with history attention and selective transformation. In IJCAI.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Deep multi-task learning for aspect term extraction with memory interaction", |
|
"authors": [ |
|
{ |
|
"first": "Xin", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wai", |
|
"middle": [], |
|
"last": "Lam", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/D17-1310" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xin Li and Wai Lam. 2017. Deep multi-task learning for aspect term extraction with memory interaction. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copen- hagen, Denmark. Association for Computational Lin- guistics.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Finegrained opinion mining with recurrent neural networks and word embeddings", |
|
"authors": [ |
|
{ |
|
"first": "Pengfei", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shafiq", |
|
"middle": [], |
|
"last": "Joty", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Helen", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 2015 conference on empirical methods in natural language processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1433--1443", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Pengfei Liu, Shafiq Joty, and Helen Meng. 2015. Fine- grained opinion mining with recurrent neural net- works and word embeddings. In Proceedings of the 2015 conference on empirical methods in natural language processing, pages 1433-1443.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "GRACE: Gradient harmonized and cascaded labeling for aspect-based sentiment analysis", |
|
"authors": [ |
|
{ |
|
"first": "Huaishao", |
|
"middle": [], |
|
"last": "Luo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lei", |
|
"middle": [], |
|
"last": "Ji", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tianrui", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daxin", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Nan", |
|
"middle": [], |
|
"last": "Duan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Findings of the Association for Computational Linguistics: EMNLP 2020", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "54--64", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.findings-emnlp.6" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Huaishao Luo, Lei Ji, Tianrui Li, Daxin Jiang, and Nan Duan. 2020. GRACE: Gradient harmonized and cas- caded labeling for aspect-based sentiment analysis. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 54-64, Online. As- sociation for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Exploring sequence-tosequence learning in aspect term extraction", |
|
"authors": [ |
|
{ |
|
"first": "Dehong", |
|
"middle": [], |
|
"last": "Ma", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sujian", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Fangzhao", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xing", |
|
"middle": [], |
|
"last": "Xie", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Houfeng", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "3538--3547", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/P19-1344" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dehong Ma, Sujian Li, Fangzhao Wu, Xing Xie, and Houfeng Wang. 2019. Exploring sequence-to- sequence learning in aspect term extraction. In Pro- ceedings of the 57th Annual Meeting of the Asso- ciation for Computational Linguistics, pages 3538- 3547, Florence, Italy. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "A joint training dual-mrc framework for aspect based sentiment analysis", |
|
"authors": [ |
|
{ |
|
"first": "Yue", |
|
"middle": [], |
|
"last": "Mao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yi", |
|
"middle": [], |
|
"last": "Shen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chao", |
|
"middle": [], |
|
"last": "Yu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Longjun", |
|
"middle": [], |
|
"last": "Cai", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2021, |
|
"venue": "AAAI", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yue Mao, Yi Shen, Chao Yu, and Longjun Cai. 2021. A joint training dual-mrc framework for aspect based sentiment analysis. In AAAI.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "YASO: A targeted sentiment analysis evaluation dataset for open-domain reviews", |
|
"authors": [ |
|
{ |
|
"first": "Matan", |
|
"middle": [], |
|
"last": "Orbach", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Orith", |
|
"middle": [], |
|
"last": "Toledo-Ronen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Artem", |
|
"middle": [], |
|
"last": "Spector", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ranit", |
|
"middle": [], |
|
"last": "Aharonov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yoav", |
|
"middle": [], |
|
"last": "Katz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Noam", |
|
"middle": [], |
|
"last": "Slonim", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2021, |
|
"venue": "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "9154--9173", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2021.emnlp-main.721" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Matan Orbach, Orith Toledo-Ronen, Artem Spector, Ranit Aharonov, Yoav Katz, and Noam Slonim. 2021. YASO: A targeted sentiment analysis evaluation dataset for open-domain reviews. In Proceedings of the 2021 Conference on Empirical Methods in Nat- ural Language Processing, pages 9154-9173, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Dare: Data augmented relation extraction with gpt-2", |
|
"authors": [ |
|
{ |
|
"first": "Yannis", |
|
"middle": [], |
|
"last": "Papanikolaou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrea", |
|
"middle": [], |
|
"last": "Pierleoni", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "SciNLP workshop at the Conference on Automated Knowledge Base Construction (AKBC)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yannis Papanikolaou and Andrea Pierleoni. 2020. Dare: Data augmented relation extraction with gpt-2. In SciNLP workshop at the Conference on Automated Knowledge Base Construction (AKBC).", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Data augmentation for spoken language understanding via pretrained language models", |
|
"authors": [ |
|
{ |
|
"first": "Baolin", |
|
"middle": [], |
|
"last": "Peng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chenguang", |
|
"middle": [], |
|
"last": "Zhu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Zeng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jianfeng", |
|
"middle": [], |
|
"last": "Gao", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"arXiv": [ |
|
"arXiv:2004.13952" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Baolin Peng, Chenguang Zhu, Michael Zeng, and Jian- feng Gao. 2020. Data augmentation for spoken lan- guage understanding via pretrained language models. arXiv preprint arXiv:2004.13952.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Glove: Global vectors for word representation", |
|
"authors": [ |
|
{ |
|
"first": "Jeffrey", |
|
"middle": [], |
|
"last": "Pennington", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Richard", |
|
"middle": [], |
|
"last": "Socher", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Christopher D", |
|
"middle": [], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1532--1543", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. Glove: Global vectors for word rep- resentation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pages 1532-1543.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "SemEval-2016 task 5: Aspect based sentiment analysis", |
|
"authors": [ |
|
{ |
|
"first": "Maria", |
|
"middle": [], |
|
"last": "Pontiki", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dimitris", |
|
"middle": [], |
|
"last": "Galanis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Haris", |
|
"middle": [], |
|
"last": "Papageorgiou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ion", |
|
"middle": [], |
|
"last": "Androutsopoulos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Suresh", |
|
"middle": [], |
|
"last": "Manandhar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Al-", |
|
"middle": [], |
|
"last": "Mohammad", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mahmoud", |
|
"middle": [], |
|
"last": "Smadi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yanyan", |
|
"middle": [], |
|
"last": "Al-Ayyoub", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bing", |
|
"middle": [], |
|
"last": "Zhao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Orph\u00e9e", |
|
"middle": [], |
|
"last": "Qin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V\u00e9ronique", |
|
"middle": [], |
|
"last": "De Clercq", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Marianna", |
|
"middle": [], |
|
"last": "Hoste", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xavier", |
|
"middle": [], |
|
"last": "Apidianaki", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Natalia", |
|
"middle": [], |
|
"last": "Tannier", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Evgeniy", |
|
"middle": [], |
|
"last": "Loukachevitch", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Kotelnikov", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "19--30", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/S16-1002" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Maria Pontiki, Dimitris Galanis, Haris Papageorgiou, Ion Androutsopoulos, Suresh Manandhar, Moham- mad AL-Smadi, Mahmoud Al-Ayyoub, Yanyan Zhao, Bing Qin, Orph\u00e9e De Clercq, V\u00e9ronique Hoste, Marianna Apidianaki, Xavier Tannier, Na- talia Loukachevitch, Evgeniy Kotelnikov, Nuria Bel, Salud Mar\u00eda Jim\u00e9nez-Zafra, and G\u00fcl\u015fen Eryigit. 2016. SemEval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), pages 19-30, San Diego, California. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Dimitris Galanis, Haris Papageorgiou, Suresh Manandhar, and Ion Androutsopoulos", |
|
"authors": [ |
|
{ |
|
"first": "Maria", |
|
"middle": [], |
|
"last": "Pontiki", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Maria Pontiki, Dimitris Galanis, Haris Papageorgiou, Suresh Manandhar, and Ion Androutsopoulos. 2015.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "Aspect based sentiment analysis", |
|
"authors": [], |
|
"year": 2015, |
|
"venue": "Proceedings of the 9th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "486--495", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/S15-2082" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "SemEval-2015 task 12: Aspect based sentiment anal- ysis. In Proceedings of the 9th International Work- shop on Semantic Evaluation (SemEval 2015), pages 486-495, Denver, Colorado. Association for Compu- tational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Harris Papageorgiou, Ion Androutsopoulos, and Suresh Manandhar", |
|
"authors": [ |
|
{ |
|
"first": "Maria", |
|
"middle": [], |
|
"last": "Pontiki", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dimitris", |
|
"middle": [], |
|
"last": "Galanis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "John", |
|
"middle": [], |
|
"last": "Pavlopoulos", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 8th International Workshop on Semantic Evaluation (Se-mEval 2014)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "27--35", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.3115/v1/S14-2004" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Maria Pontiki, Dimitris Galanis, John Pavlopoulos, Har- ris Papageorgiou, Ion Androutsopoulos, and Suresh Manandhar. 2014. SemEval-2014 task 4: Aspect based sentiment analysis. In Proceedings of the 8th International Workshop on Semantic Evaluation (Se- mEval 2014), pages 27-35, Dublin, Ireland. Associa- tion for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Unleash GPT-2 power for event detection", |
|
"authors": [], |
|
"year": null, |
|
"venue": "Proceedings of the 59th", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2021.acl-long.490" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Amir Pouran Ben Veyseh, Viet Lai, Franck Dernon- court, and Thien Huu Nguyen. 2021. Unleash GPT-2 power for event detection. In Proceedings of the 59th", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing", |
|
"authors": [], |
|
"year": null, |
|
"venue": "", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "6271--6282", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer- ence on Natural Language Processing (Volume 1: Long Papers), pages 6271-6282, Online. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "Franck Dernoncourt, Dejing Dou, and Thien Huu Nguyen. 2020a. Introducing syntactic structures into target opinion word extraction with deep learning", |
|
"authors": [ |
|
{ |
|
"first": "Ben", |
|
"middle": [], |
|
"last": "Amir Pouran", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Nasim", |
|
"middle": [], |
|
"last": "Veyseh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Nouri", |
|
"suffix": "" |
|
} |
|
], |
|
"year": null, |
|
"venue": "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "8947--8956", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.emnlp-main.719" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Amir Pouran Ben Veyseh, Nasim Nouri, Franck Dernon- court, Dejing Dou, and Thien Huu Nguyen. 2020a. Introducing syntactic structures into target opinion word extraction with deep learning. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 8947- 8956, Online. Association for Computational Lin- guistics.", |
|
"links": null |
|
}, |
|
"BIBREF30": { |
|
"ref_id": "b30", |
|
"title": "Improving aspect-based sentiment analysis with gated graph convolutional networks and syntax-based regulation", |
|
"authors": [], |
|
"year": null, |
|
"venue": "Findings of the Association for Computational Linguistics: EMNLP 2020", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "4543--4548", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.findings-emnlp.407" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Amir Pouran Ben Veyseh, Nasim Nouri, Franck Dernon- court, Quan Hung Tran, Dejing Dou, and Thien Huu Nguyen. 2020b. Improving aspect-based sentiment analysis with gated graph convolutional networks and syntax-based regulation. In Findings of the Associ- ation for Computational Linguistics: EMNLP 2020, pages 4543-4548, Online. Association for Computa- tional Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF31": { |
|
"ref_id": "b31", |
|
"title": "Language models are unsupervised multitask learners", |
|
"authors": [ |
|
{ |
|
"first": "Alec", |
|
"middle": [], |
|
"last": "Radford", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jeffrey", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rewon", |
|
"middle": [], |
|
"last": "Child", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "David", |
|
"middle": [], |
|
"last": "Luan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dario", |
|
"middle": [], |
|
"last": "Amodei", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ilya", |
|
"middle": [], |
|
"last": "Sutskever", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "OpenAI blog", |
|
"volume": "1", |
|
"issue": "8", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.", |
|
"links": null |
|
}, |
|
"BIBREF32": { |
|
"ref_id": "b32", |
|
"title": "Exploring the limits of transfer learning with a unified text-to-text transformer", |
|
"authors": [ |
|
{ |
|
"first": "Colin", |
|
"middle": [], |
|
"last": "Raffel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Noam", |
|
"middle": [], |
|
"last": "Shazeer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Adam", |
|
"middle": [], |
|
"last": "Roberts", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Katherine", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sharan", |
|
"middle": [], |
|
"last": "Narang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Matena", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yanqi", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wei", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Peter J", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"arXiv": [ |
|
"arXiv:1910.10683" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2019. Exploring the limits of transfer learning with a unified text-to-text trans- former. arXiv preprint arXiv:1910.10683.", |
|
"links": null |
|
}, |
|
"BIBREF33": { |
|
"ref_id": "b33", |
|
"title": "EliXa: A modular and flexible ABSA platform", |
|
"authors": [ |
|
{ |
|
"first": "Xabier", |
|
"middle": [], |
|
"last": "I\u00f1aki San Vicente", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rodrigo", |
|
"middle": [], |
|
"last": "Saralegi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Agerri", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 9th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "748--752", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/S15-2127" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "I\u00f1aki San Vicente, Xabier Saralegi, and Rodrigo Agerri. 2015. EliXa: A modular and flexible ABSA plat- form. In Proceedings of the 9th International Work- shop on Semantic Evaluation (SemEval 2015), pages 748-752, Denver, Colorado. Association for Compu- tational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF34": { |
|
"ref_id": "b34", |
|
"title": "NLANGP at SemEval-2016 task 5: Improving aspect based sentiment analysis using neural network features", |
|
"authors": [ |
|
{ |
|
"first": "Zhiqiang", |
|
"middle": [], |
|
"last": "Toh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jian", |
|
"middle": [], |
|
"last": "Su", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "282--288", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/S16-1045" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhiqiang Toh and Jian Su. 2016. NLANGP at SemEval- 2016 task 5: Improving aspect based sentiment anal- ysis using neural network features. In Proceedings of the 10th International Workshop on Semantic Eval- uation (SemEval-2016), pages 282-288, San Diego, California. Association for Computational Linguis- tics.", |
|
"links": null |
|
}, |
|
"BIBREF35": { |
|
"ref_id": "b35", |
|
"title": "DLIREC: Aspect term extraction and term polarity classification system", |
|
"authors": [ |
|
{ |
|
"first": "Zhiqiang", |
|
"middle": [], |
|
"last": "Toh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wenting", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 8th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "235--240", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.3115/v1/S14-2038" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhiqiang Toh and Wenting Wang. 2014. DLIREC: As- pect term extraction and term polarity classification system. In Proceedings of the 8th International Work- shop on Semantic Evaluation (SemEval 2014), pages 235-240, Dublin, Ireland. Association for Computa- tional Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF36": { |
|
"ref_id": "b36", |
|
"title": "Cider: Consensus-based image description evaluation", |
|
"authors": [ |
|
{ |
|
"first": "Ramakrishna", |
|
"middle": [], |
|
"last": "Vedantam", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lawrence", |
|
"middle": [], |
|
"last": "Zitnick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Devi", |
|
"middle": [], |
|
"last": "Parikh", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the IEEE conference on computer vision and pattern recognition", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "4566--4575", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. 2015. Cider: Consensus-based image de- scription evaluation. In Proceedings of the IEEE conference on computer vision and pattern recogni- tion, pages 4566-4575.", |
|
"links": null |
|
}, |
|
"BIBREF37": { |
|
"ref_id": "b37", |
|
"title": "Human-like decision making: Document-level aspect sentiment classification via hierarchical reinforcement learning", |
|
"authors": [ |
|
{ |
|
"first": "Jingjing", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Changlong", |
|
"middle": [], |
|
"last": "Sun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shoushan", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jiancheng", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luo", |
|
"middle": [], |
|
"last": "Si", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Min", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaozhong", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Guodong", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "5581--5590", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/D19-1560" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jingjing Wang, Changlong Sun, Shoushan Li, Jiancheng Wang, Luo Si, Min Zhang, Xiaozhong Liu, and Guodong Zhou. 2019. Human-like decision making: Document-level aspect sentiment classification via hierarchical reinforcement learning. In Proceedings of the 2019 Conference on Empirical Methods in Nat- ural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5581-5590, Hong Kong, China. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF38": { |
|
"ref_id": "b38", |
|
"title": "Recursive neural conditional random fields for aspect-based sentiment analysis", |
|
"authors": [ |
|
{ |
|
"first": "Wenya", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Sinno Jialin Pan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaokui", |
|
"middle": [], |
|
"last": "Dahlmeier", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Xiao", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "616--626", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/D16-1059" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wenya Wang, Sinno Jialin Pan, Daniel Dahlmeier, and Xiaokui Xiao. 2016. Recursive neural conditional random fields for aspect-based sentiment analysis. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 616- 626, Austin, Texas. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF39": { |
|
"ref_id": "b39", |
|
"title": "Simple statistical gradientfollowing algorithms for connectionist reinforcement learning", |
|
"authors": [ |
|
{ |
|
"first": "Ronald", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Williams", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1992, |
|
"venue": "Kluwer Academic", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ronald J. Williams. 1992. Simple statistical gradient- following algorithms for connectionist reinforcement learning. In Kluwer Academic.", |
|
"links": null |
|
}, |
|
"BIBREF40": { |
|
"ref_id": "b40", |
|
"title": "Phrase dependency parsing for opinion mining", |
|
"authors": [ |
|
{ |
|
"first": "Yuanbin", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Qi", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xuanjing", |
|
"middle": [], |
|
"last": "Huang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lide", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1533--1541", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yuanbin Wu, Qi Zhang, Xuanjing Huang, and Lide Wu. 2009. Phrase dependency parsing for opinion mining. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, pages 1533-1541, Singapore. Association for Com- putational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF41": { |
|
"ref_id": "b41", |
|
"title": "Double embeddings and CNN-based sequence labeling for aspect extraction", |
|
"authors": [ |
|
{ |
|
"first": "Hu", |
|
"middle": [], |
|
"last": "Xu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bing", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lei", |
|
"middle": [], |
|
"last": "Shu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Philip", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Yu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "592--598", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/P18-2094" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hu Xu, Bing Liu, Lei Shu, and Philip S. Yu. 2018. Dou- ble embeddings and CNN-based sequence labeling for aspect extraction. In Proceedings of the 56th An- nual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 592-598, Melbourne, Australia. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF42": { |
|
"ref_id": "b42", |
|
"title": "Constituency lattice encoding for aspect term extraction", |
|
"authors": [ |
|
{ |
|
"first": "Yunyi", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kun", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaojun", |
|
"middle": [], |
|
"last": "Quan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Weizhou", |
|
"middle": [], |
|
"last": "Shen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Qinliang", |
|
"middle": [], |
|
"last": "Su", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 28th International Conference on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "844--855", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.coling-main.73" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yunyi Yang, Kun Li, Xiaojun Quan, Weizhou Shen, and Qinliang Su. 2020. Constituency lattice encoding for aspect term extraction. In Proceedings of the 28th International Conference on Computational Lin- guistics, pages 844-855, Barcelona, Spain (Online). International Committee on Computational Linguis- tics.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF1": { |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Statistics of the SemEval datasets", |
|
"type_str": "table", |
|
"num": null |
|
}, |
|
"TABREF3": { |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "F1 scores on the test sets of the SemEval datasets. The proposed model ATEOE is significantly better than prior work (p < 0.05).", |
|
"type_str": "table", |
|
"num": null |
|
}, |
|
"TABREF5": { |
|
"html": null, |
|
"content": "<table><tr><td>erate sentences that are similar to the complement</td></tr><tr><td>sentences S \u2032 .</td></tr></table>", |
|
"text": "Performance of the ablated models on test sets.", |
|
"type_str": "table", |
|
"num": null |
|
}, |
|
"TABREF6": { |
|
"html": null, |
|
"content": "<table><tr><td>Language Model Lap14 Res14 Res15 Res16</td></tr><tr><td>GPT-2</td></tr></table>", |
|
"text": "86.71 88.99 75.41 82.58 BART 84.32 88.05 74.91 79.49 T5 84.18 86.95 73.16 79.15", |
|
"type_str": "table", |
|
"num": null |
|
}, |
|
"TABREF7": { |
|
"html": null, |
|
"content": "<table><tr><td colspan=\"5\">Language Model BLUE-4 METEOR ROUGE-1 ROUGE-2</td></tr><tr><td>GPT-2</td><td>12.05</td><td>12.25</td><td>31.89</td><td>10.33</td></tr><tr><td>BART</td><td>9.39</td><td>10.14</td><td>29.05</td><td>9.06</td></tr><tr><td>T5</td><td>13.10</td><td>11.92</td><td>30.33</td><td>8.95</td></tr></table>", |
|
"text": "Performance of the final ATE model on test sets with different language models in step III.", |
|
"type_str": "table", |
|
"num": null |
|
}, |
|
"TABREF8": { |
|
"html": null, |
|
"content": "<table><tr><td>Model</td><td>Lap14 Res14 Res15 Res16</td></tr><tr><td>DE-CNN</td><td>74.37</td></tr></table>", |
|
"text": "77.61 70.00 70.68 SoftProto 79.85 82.22 76.80 70.93 ATEOA (ours) 81.92 83.69 77.39 73.49", |
|
"type_str": "table", |
|
"num": null |
|
}, |
|
"TABREF9": { |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Performance for tail aspect terms on test data.", |
|
"type_str": "table", |
|
"num": null |
|
} |
|
} |
|
} |
|
} |