{"id": "test_0", "sentence1": "On one hand, assuming that the parallel sentences have the same meaning, many datasets find the aligned sentences to have higher string overlap (as measured by BLEU).", "sentence2": "the two sentences should have different styles, and may vary a lot in expressions: and thus leading to a lower BLEU.", "label": "contrasting"} {"id": "test_1", "sentence1": "Recent solutions (Gidaris and Komodakis, 2018) leverage a memory component to maintain models' learning experience, e.g., by finding from a supervised stage the content that is similar to the unseen classes, leading to the state-of-the-art performance.", "sentence2": "the memory weights are static during inference and the capability of the model is still limited when adapted to new classes.", "label": "contrasting"} {"id": "test_2", "sentence1": "A larger KES leads to predict more unique keyphrases, append less absolutely incorrect keyphrases and improve the chance to output more accurate keyphrases.", "sentence2": "generating more unique keyphrases may also lead to more incorrect predictions, which will degrade the F 1 @M scores since F 1 @M considers all the unique predictions without a fixed cutoff.", "label": "contrasting"} {"id": "test_3", "sentence1": "For instance, all the baselines produce the keyphrase \"debugging\" at least three times.", "sentence2": "our ExHiRD-h only generates it once, which demonstrates that our proposed method is more powerful in avoiding duplicated keyphrases.", "label": "contrasting"} {"id": "test_4", "sentence1": "Those models mainly focus on improving decoders based on the constraint of hierarchical paths.", "sentence2": "we propose an effective hierarchy-aware global model, HiAGM, that extracts label-wise text features with hierarchy encoders based on prior hierarchy information.", "label": "contrasting"} {"id": "test_5", "sentence1": "Previous global models classify labels upon the original textual information and improve the decoder with predefined hierarchy paths.", "sentence2": "we construct a novel end-to-end hierarchy-aware global model (HiAGM) for the mutual interaction of text features and label correlations.", "label": "contrasting"} {"id": "test_6", "sentence1": "As for the amount of operators' effort, we observed a slight decrease in HTER with the increase of pre-filtering conditions, indicating an improvement in the quality of candidates.", "sentence2": "hTER scores were all between 0.1 and 0.2, much below the 0.4 acceptability threshold defined by Turchi et al.", "label": "contrasting"} {"id": "test_7", "sentence1": "Finally, we observe that despite reducing the ouput diversity and novelty, the reduction of expert effort by Reviewer\u22652 in terms of the percentage of the obtained pairs is not attainable by a machine yet.", "sentence2": "automatic filtering (Reviewer machine ) is a viable solution since (i) it helps the NGO operators save time better than human filter \u22651, (ii) it preserves diversity and novelty better than Reviewer\u22652 and in line with Reviewer\u22651 .", "label": "contrasting"} {"id": "test_8", "sentence1": "In this scenario, automation strategies, such as natural language generation, are necessary to help NGO operators in their countering effort.", "sentence2": "these automation approaches are not mature yet, since they suffer from the lack of sufficient amount of quality data and tend to produce generic/repetitive responses.", "label": "contrasting"} {"id": "test_9", "sentence1": "However, gold data for the target language (stage) is usually inaccessible, often preventing evaluation against human judgment.", "sentence2": "we here propose several alternative evaluation set-ups as an integral part of our methodology.", "label": "contrasting"} {"id": "test_10", "sentence1": "Words, such as German Sonnenschein for which a translational equivalent already exists in the Source (\"sunshine\"; see Figure 1), mainly rely on translation, while the prediction step acts as an optional refinement procedure.", "sentence2": "the prediction step is crucial for words, such as Erdbeben, whose translational equivalents (\"earthquake\") are missing in the Source.", "label": "contrasting"} {"id": "test_11", "sentence1": "We want to point out that not every single entry should be considered meaningful because of noise in the embedding vocabulary caused by typos and tokenization errors.", "sentence2": "choosing the \"best\" size for an emotion lexicon necessarily translates into a quality-coverage trade-off for which there is no general solution.", "label": "contrasting"} {"id": "test_12", "sentence1": "Neural MT (NMT) approaches have certainly made significant progress in this direction.", "sentence2": "the diversity of possible outcomes makes it harder to evaluate MT models.", "label": "contrasting"} {"id": "test_13", "sentence1": "Dreyer and Marcu (2012) showed that if multiple human translations are used, any automatic MT evaluation metric achieves a substantially higher correlation with human judgments.", "sentence2": "multiple translations are hardly ever available in practice due to the cost of collecting them.", "label": "contrasting"} {"id": "test_14", "sentence1": "The distinction between intended and perceived sarcasm, also referred to as encoded and decoded sarcasm, respectively, has been pointed out in previous research (Kaufer, 1981;Rockwell and Theriot, 2001).", "sentence2": "it has not been considered in a computational context thus far when building datasets for textual sarcasm detection.", "label": "contrasting"} {"id": "test_15", "sentence1": "Text summarization has recently received increased attention with the rise of deep learning-based endto-end models, both for extractive and abstractive variants.", "sentence2": "so far, only single-document summarization has profited from this trend.", "label": "contrasting"} {"id": "test_16", "sentence1": "Apparently, a summarization method is desirable to achieve a ROUGE score of 100, i.e., a system output is identical to the reference.", "sentence2": "this is an unrealistic goal for the task setting on the Gigaword dataset.", "label": "contrasting"} {"id": "test_17", "sentence1": "They addressed the problem where an abstractive model made mistakes in facts (e.g., tuples of subjects, predicates, and objects).", "sentence2": "we also regularly see examples where the abstractive model generates unexpected words.", "label": "contrasting"} {"id": "test_18", "sentence1": "For JNC, we use the pretrained BERT model for Japanese text (Kikuta, 2019).", "sentence2": "no large-scale Japanese corpus for semantic inference (counterpart to MultiNLI) is available.", "label": "contrasting"} {"id": "test_19", "sentence1": "We could confirm the improvements from the support scores, entailment ratio, and human judgments.", "sentence2": "the ROUGE scores between system and reference headlines did not indicate a clear difference.", "label": "contrasting"} {"id": "test_20", "sentence1": "Likert-score based self-reported user rating is widely adopted by social conversational systems, such as Amazon Alexa Prize chatbots.", "sentence2": "selfreported user rating suffers from bias and variance among different users.", "label": "contrasting"} {"id": "test_21", "sentence1": "ELMo (Peters et al., 2018), BERT (Devlin et al., 2019), and XLNet (Yang et al., 2019) have achieved great success in many NLP tasks.", "sentence2": "it is difficult to apply them in the industrial dialog system due to their low computational efficiency.", "label": "contrasting"} {"id": "test_22", "sentence1": "Previous work has demonstrated that neural encoders capture a rich hierarchy of syntactic and semantic information (Jawahar et al., 2019;Clark et al., 2019).", "sentence2": "reasoning capability and commonsense knowledge are not captured sufficiently (Young et al., 2018).", "label": "contrasting"} {"id": "test_23", "sentence1": "BERT-MC and RoBERTa-MC obtain similar results with BERT and RoBERTa, respectively.", "sentence2": "even RoBERTa is far behind human performance 23 points on R@1, indicating that MuTual is indeed a challenging dataset, which opens the door for tackling new and complex reasoning problems in multi-turn conversations.", "label": "contrasting"} {"id": "test_24", "sentence1": "The language score is evaluated individually, without considering the discourse coherence.", "sentence2": "a reasonable response should establish links in meaning with context, which is also an important aspect of humanlike responses.", "label": "contrasting"} {"id": "test_25", "sentence1": "In practice, the solver can always find a solution to linearize the subtrees with the constraints.", "sentence2": "it sometimes cannot find any solution to directly linearize the full tree within the time limit (1-10% of the cases depending on the treebank), because there are more nodes and more constraints in the full tree.", "label": "contrasting"} {"id": "test_26", "sentence1": "However, these approaches only consider sentence-level QG.", "sentence2": "our work focus on the challenge of generating deep questions with multi-hop reasoning over document-level contexts.", "label": "contrasting"} {"id": "test_27", "sentence1": "Among them, \"Semantic Error\", \"Redundant\", and \"Unanswerable\" are noticeable errors for all models.", "sentence2": "we find that baselines have more unreasonable subject-predicate-object collocations (semantic errors) than our model.", "label": "contrasting"} {"id": "test_28", "sentence1": "Extracting relational triples from unstructured text is crucial for large-scale knowledge graph construction.", "sentence2": "few existing works excel in solving the overlapping triple problem where multiple relational triples in the same sentence share the same entities.", "label": "contrasting"} {"id": "test_29", "sentence1": "To reduce manual work, recent studies have investigated neural network-based methods, which deliver state-of-the-art performance.", "sentence2": "most existing neural models like (Miwa and Bansal, 2016) achieve joint learning of entities and relations only through parameter sharing but not joint decoding.", "label": "contrasting"} {"id": "test_30", "sentence1": "Therefore, the object tagger for relation \"Work in\" will not identify the span of \"Washington\", i.e., the output of both start and end position are all zeros as shown in Figure 2.", "sentence2": "the relation \"Birth place\" holds between \"Jackie R. Brown\" and \"Washington\", so the corresponding object tagger outputs the span of the candidate object \"Washington\".", "label": "contrasting"} {"id": "test_31", "sentence1": "Such inconsistent data distribution of two datasets leads to a comparatively better performance on NYT and a worse performance on WebNLG for all the baselines, exposing their drawbacks in extracting overlapping relational triples.", "sentence2": "the CASREL model and its variants (i.e., CASREL random and CAS-REL LST M ) all achieve a stable and competitive performance on both NYT and WebNLG datasets, demonstrating the effectiveness of the proposed framework in solving the overlapping problem.", "label": "contrasting"} {"id": "test_32", "sentence1": "In other words, it implies that identifying relations is somehow easier than identifying entities for our model.", "sentence2": "to NYT, for WebNLG, the performance gap between (E1, E2) and (E1, R, E2) is comparatively larger than that between (E1, R, E2) and (E1, R)/(R, E2).", "label": "contrasting"} {"id": "test_33", "sentence1": "We also find that there is only a trivial gap between the F1-score on (E1, E2) and (E1, R, E2), but an obvious gap between (E1, R, E2) and (E1, R)/(R, E2).", "sentence2": "it reveals that most relations for the entity pairs in extracted triples are correctly identified while some extracted entities fail to form a valid relational triple", "label": "contrasting"} {"id": "test_34", "sentence1": "Information Extraction (IE), and specifically Relation Extraction (RE), can improve the access to central information for downstream tasks (Santos et al., 2015;Zeng et al., 2014;Jiang et al., 2016;Miwa and Bansal, 2016;Luan et al., 2018a).", "sentence2": "the focus of current RE systems and datasets is either too narrow, i.e., a handful of semantic relations, such as 'USED-FOR' and 'SYNONYMY', or too broad, i.e., an unbounded number of generic relations extracted from large, heterogeneous corpora (Niklaus et al., 2018), referred to as Open IE (OIE) (Etzioni et al., 2005;Banko et al., 2007).", "label": "contrasting"} {"id": "test_35", "sentence1": "It has been shown that scientific texts contain many unique relation types and, therefore, it is not feasible to create separate narrow IE classifiers for these (Groth et al., 2018).", "sentence2": "oIE systems are primarily developed for the Web and news-wire domain and have been shown to perform poorly on scientific texts.", "label": "contrasting"} {"id": "test_36", "sentence1": "Lei et al. (2017) conduct word pair interaction score to capture both linear and quadratic relation for argument representation.", "sentence2": "these methods utilize the pre-trained embeddings for mining the interaction features and ignore the geometric structure information entailed in discourse arguments and their relation.", "label": "contrasting"} {"id": "test_37", "sentence1": "Xu et al. (2019) propose a topic tensor network (TTN) to model the sentence-level interactions and topic-level rel\u0002evance among arguments for this task.", "sentence2": "few studies model discourse relations by translating them in the low-dimensional embedding space as we do in this work.", "label": "contrasting"} {"id": "test_38", "sentence1": "With the increasing of the number of encoder layers, the model could capture the richer semantic information.", "sentence2": "the results imply that with the more encoder layers considered, the model could incur the over-fitting problem due to adding more parameters.", "label": "contrasting"} {"id": "test_39", "sentence1": "Based on BLEU (Papineni et al., 2002) scores, previous work (Belinkov et al., 2017) suggests that translating into morphologically rich languages, such as Hungarian or Finnish, is harder than translating into morphologically poor ones, such as English.", "sentence2": "a major obstacle in the crosslingual comparison of MT systems is that many automatic evaluation metrics, including BLEU and METEOR (Banerjee and Lavie, 2005), are not cross-lingually comparable.", "label": "contrasting"} {"id": "test_40", "sentence1": "Reliable reference-free evaluation metrics, directly measuring the (semantic) correspondence between the source language text and system translation, would remove the need for human references and allow for unlimited MT evaluations: any monolingual corpus could be used for evaluating MT systems.", "sentence2": "the proposals of referencefree MT evaluation metrics have been few and far apart and have required either non-negligible supervision (i.e., human translation quality labels) (Specia et al., 2010) or language-specific preprocessing like semantic parsing (Lo et al., 2014;Lo, 2019), both hindering the wide applicability of the proposed metrics.", "label": "contrasting"} {"id": "test_41", "sentence1": "Position encoding (PE), an essential part of self-attention networks (SANs), is used to preserve the word order information for natural language processing tasks, generating fixed position indices for input sequences.", "sentence2": "in cross-lingual scenarios, e.g., machine translation, the PEs of source and target sentences are modeled independently.", "label": "contrasting"} {"id": "test_42", "sentence1": "The filtering step as performed by Grave et al. (2018) consisted in only keeping the lines exceeding 100 bytes in length.", "sentence2": "considering that Common Crawl is a mutilingual UTF-8 encoded corpus, this 100-byte threshold creates a huge disparity between ASCII and non-ASCII encoded languages.", "label": "contrasting"} {"id": "test_43", "sentence1": "Considering the discussion above, we believe an interesting follow-up to our experiments would be training the ELMo models for more of the languages included in the OSCAR corpus.", "sentence2": "training ELMo is computationally costly, and one way to estimate this cost, as pointed out by Strubell et al. (2019), is by using the training times of each model to compute both power consumption and CO2 emissions.", "label": "contrasting"} {"id": "test_44", "sentence1": "Even though it would have been interesting to replicate all our experiments and computational cost estimations with state-of-the-art fine-tuning models such as BERT, XLNet, RoBERTa or AL-BERT, we recall that these transformer-based architectures are extremely costly to train, as noted by the BERT authors on the official BERT GitHub repository, 21 and are currently beyond the scope of our computational infrastructure.", "sentence2": "we believe that ELMo contextualized word embeddings remain a useful model that still provide an extremely good trade-off between performance to training cost, even setting new state-of-the-art scores in parsing and POS tagging for our five chosen languages, performing even better than the multilingual mBERT model.", "label": "contrasting"} {"id": "test_45", "sentence1": "All neural models achieve a SG score significantly greater than a random baseline (dashed line).", "sentence2": "the range within neural models is notable, with the bestperforming model (GPT-2-XL) scoring over twice as high as the worst-performing model (LSTM).", "label": "contrasting"} {"id": "test_46", "sentence1": "The TTG ap-proach cannot make use at run-time of an En QE model without translating the caption back to English and thus again requiring perfect translation in order not to ruin the predicted quality score.", "sentence2": "the PLuGS approach appears to be best suited for leveraging an existing En QE model, due to the availability of the generated bilingual output that tends to maintain consistency between the generated EN-& X-language outputs, with respect to accuracy; therefore, directly applying an English QE model appears to be the most appropriate scalable solution.", "label": "contrasting"} {"id": "test_47", "sentence1": "They read one sentence at a time and provided a suspense judgement using the fivepoint scale consisting of Big Decrease in suspense (1% of the cases), Decrease (11%), Same (50%), Increase (31%), and Big Increase (7%).", "sentence2": "to prior work (Delatorre et al., 2018), a relative rather than absolute scale was used.", "label": "contrasting"} {"id": "test_48", "sentence1": "One of our main hypotheses is that modeling edit sequences is better suited for this task than generating comments from scratch.", "sentence2": "a counter argument could be that a comment generation model could be trained from substantially more data, since it is much easier to obtain parallel data in the form (method, comment), without the constraints of simultaneous code/comment edits.", "label": "contrasting"} {"id": "test_49", "sentence1": "Most of these approaches focus on code summarization or comment generation which only require single code-NL pairs for training and evaluation as the task entails generating a natural language summary of a given code snippet.", "sentence2": "our proposed task requires two code-NL pairs that are assumed to hold specific parallel relationships with one another.", "label": "contrasting"} {"id": "test_50", "sentence1": "FACTEDITOR has a larger number of correct editing (CQT) than ENCDECEDITOR for fact-based text editing.", "sentence2": "eNCDeCeDITOR often includes a larger number of unnecessary rephrasings (UPARA) than FACTeDITOR.", "label": "contrasting"} {"id": "test_51", "sentence1": "ENCDECEDITOR often generates the same facts multiple times (RPT) or facts in different relations (DREL).", "sentence2": "fACTE-DITOR can seldomly make such errors.", "label": "contrasting"} {"id": "test_52", "sentence1": "ENCDECEDITOR cannot effectively eliminate the description about an unsupported fact (in orange) appearing in the draft text.", "sentence2": "fACTEDI-TOR can deal with the problem well.", "label": "contrasting"} {"id": "test_53", "sentence1": "Gehrmann et al. (2018) utilized a data-efficient content selector, by aligning source and target, to restrict the model\u2019s attention to likely-to-copy phrases.", "sentence2": "we use the content selector to find domain knowledge alignment between source and target.", "label": "contrasting"} {"id": "test_54", "sentence1": "Such data-driven approaches achieve good performance on several benchmarks like E2E challenge (Novikova et al., 2017), WebNLG challenge (Gardent et al., 2017) and WIKIBIO (Lebret et al., 2016).", "sentence2": "they rely on massive amount of training data.", "label": "contrasting"} {"id": "test_55", "sentence1": "Ma et al. (2019) propose low-resource table-to-text generation with 1,000 paired examples and large-scale target-side examples.", "sentence2": "in our setting, only tens to hundreds of paired training examples are required, meanwhile without the need for any target examples.", "label": "contrasting"} {"id": "test_56", "sentence1": "This spillover is potentially sensitive only to Levenshtein distance.", "sentence2": "confusability is sensitive to fine-grained perceptual structure.", "label": "contrasting"} {"id": "test_57", "sentence1": "For this case, IG, IC, and subjectivity all have overlapping confidence intervals, so we conclude that there is no evidence that one is better than the other.", "sentence2": "we do have evidence that IG and IC are more accurate than PMI when estimated based on clusters.", "label": "contrasting"} {"id": "test_58", "sentence1": "The idea of adding the NOTA option to a candidate set is also widely used in other language technology fields like speaker verification (Pathak and Raj, 2013).", "sentence2": "the effect of adding NOTA is rarely introduced in dialog retrieval research problems.", "label": "contrasting"} {"id": "test_59", "sentence1": "D-GPT gets the wrong answer 18% of the time (option a and c), because the input answer predicted by the CoQA baseline is also incorrect 17% of the time.", "sentence2": "with oracle answers, it is able to generate correct responses 77% of the times (option e).", "label": "contrasting"} {"id": "test_60", "sentence1": "Question Answering Using crowd-sourcing methods to create QA datasets (Rajpurkar et al., 2016;Bajaj et al., 2016;Rajpurkar et al., 2018), conversational datasets (Dinan et al., 2018), and ConvQA datasets (Choi et al., 2018;Reddy et al., 2019;Elgohary et al., 2018;Saha et al., 2018) has largely driven recent methodological advances.", "sentence2": "models trained on these ConvQA datasets typically select exact answer spans instead of generating them (Yatskar, 2019b).", "label": "contrasting"} {"id": "test_61", "sentence1": "As shown in Table 1, the hidden dimension of each building block is only 128.", "sentence2": "we introduce two linear transformations for each building block to adjust its input and output dimensions to 512.", "label": "contrasting"} {"id": "test_62", "sentence1": "One may either only use the bottlenecks for MobileBERT (correspondingly the teacher becomes BERT LARGE ) or only the invertedbottlenecks for IB-BERT (then there is no bottleneck in MobileBERT) to align their feature maps.", "sentence2": "when using both of them, we can allow IB-BERT LARGE to preserve the performance of BERT LARGE while having MobileBERT sufficiently compact.", "label": "contrasting"} {"id": "test_63", "sentence1": "This has enabled ever-increasing performance on benchmark data sets.", "sentence2": "one thing has remained relatively constant: the softmax of a dot product as the output layer.", "label": "contrasting"} {"id": "test_64", "sentence1": "Recently Graph Neural Network (GNN) has shown to be powerful in successfully tackling many tasks.", "sentence2": "there has been no attempt to exploit GNN to create taxonomies.", "label": "contrasting"} {"id": "test_65", "sentence1": "We get better performance if we tune the thresholds.", "sentence2": "we chose a harder task and proved our model has better performance than others even we simply use 0.5 as the threshold.", "label": "contrasting"} {"id": "test_66", "sentence1": "The model identifies words related to community (\"kids,\" \"neighborhood,\" \"we\") as strong negative signals for depression, supporting that depressed language reflects detachment from community.", "sentence2": "the model only focuses on these semantic themes in responses to generic backchannel categories.", "label": "contrasting"} {"id": "test_67", "sentence1": "Thanks to the increased complexity of deep neural networks and use of knowledge transfer from the language models pre\u0002trained on large-scale corpora (Peters et al., 2018; Devlin et al., 2019; Dong et al., 2019), the state of-the-art QA models have achieved human-level performance on several benchmark datasets (Rajpurkar et al., 2016, 2018)", "sentence2": "what is also crucial to the success of the recent data-driven mod\u0002els, is the availability of large-scale QA datasets", "label": "contrasting"} {"id": "test_68", "sentence1": "Some of the recent works resort to semi-supervised learning, by leveraging large amount of unlabeled text (e.g. Wikipedia) to generate synthetic QA pairs with the help of QG systems (Tang et al., 2017; Yang et al., 2017; Tang et al., 2018; Sachan and Xing, 2018).", "sentence2": "existing QG systems have overlooked an important point that generating QA pairs from a context consisting of unstructured texts, is essentially a one-to-many problem.", "label": "contrasting"} {"id": "test_69", "sentence1": "They should be semantically consistent, such that it is possible to predict the answer given the question and the context.", "sentence2": "neural QG or QAG models often generate questions irrelevant to the context and the answer (Zhang and Bansal, 2019) due to the lack of the mechanism enforcing this consistency.", "label": "contrasting"} {"id": "test_70", "sentence1": "For example, when both models are trained with 1% of the Yelp dataset, the accuracy gap is around 9%.", "sentence2": "as we increases the amount of training data to 90%, the accuracy gap drops to within 2%.", "label": "contrasting"} {"id": "test_71", "sentence1": "Upon further investigation, we find that experiments which use probabilities with image based features have an inter-quartile range of 0.05 and 0.1 for EBG and BLOG respectively whereas for experiments using probabilities with binning based features, this range is 0.32 for both datasets.", "sentence2": "inter-quartile range for exper-iments using ranks with image based features is 0.08 and 0.05 for EBG and BLOG whereas for experiments using ranks with binning based features, this range is 0.49 and 0.42 respectively.", "label": "contrasting"} {"id": "test_72", "sentence1": "If we have a large number of training samples, the architecture is capable of learning how to discriminate correctly between classes only with the original training data.", "sentence2": "in less-resourced scenarios, our proposed approaches with external knowledge integration could achieve a high positive impact.", "label": "contrasting"} {"id": "test_73", "sentence1": "These offer obvious benefits to users in terms of immediacy, interaction and convenience.", "sentence2": "it remains challenging for application providers to assess language content collected through these means.", "label": "contrasting"} {"id": "test_74", "sentence1": "We make our code publicly available for others to use for benchmarking and replication experiments.", "sentence2": "to feature-based scoring, we instead train neural networks on ASR transcriptions which are labeled with proficiency scores assigned by human examiners, and guide the networks with objectives that prioritize language understanding.", "label": "contrasting"} {"id": "test_75", "sentence1": "These results were not significantly better than the single-task POS prediction model, though we did not explore tuning the alpha weighting values for the combination models.", "sentence2": "bERT only receives a significant improvement in grading ability when using the L1 prediction task.", "label": "contrasting"} {"id": "test_76", "sentence1": "Figure 4 shows, as expected, that training a speech grader with data from an ASR system with lower word error rates produces better results.", "sentence2": "it is interesting to note that this holds true even when evaluating with data from inferior ASR systems.", "label": "contrasting"} {"id": "test_77", "sentence1": "This is because SciBERT, like other pretrained language models, is trained via language modeling objectives, which only predict words or sentences given their in-document, nearby textual context.", "sentence2": "we propose to incorporate citations into the model as a signal of inter-document relatedness, while still leveraging the model's existing strength in modeling language.", "label": "contrasting"} {"id": "test_78", "sentence1": "The candidate program should adhere to the grammatical specification of the target language.", "sentence2": "since incorporating the complete set of C++ grammatical constraints would require significant engineering effort, we instead restrict our attention to the set of \"primary expressions\" consisting of high-level control structures such as if, else, for loops, function declarations, etc.", "label": "contrasting"} {"id": "test_79", "sentence1": "For example, when there is only one statement within an if statement, the programmer can optionally include a curly brace.", "sentence2": "the pseudocode does not contain such detailed information about style.", "label": "contrasting"} {"id": "test_80", "sentence1": "This error can be ruled out by SymTable constraint if variable A is undeclared.", "sentence2": "symTable constraints do not preclude all errors related to declarations.", "label": "contrasting"} {"id": "test_81", "sentence1": "Prior works regarded SQG as a dialog generation task and recurrently produced each question.", "sentence2": "they suffered from problems caused by error cascades and could only capture limited context dependencies.", "label": "contrasting"} {"id": "test_82", "sentence1": "We expect that this category is rare because the premise is not text.", "sentence2": "since there are some textual elements in the tables, the hypothesis could paraphrase them.", "label": "contrasting"} {"id": "test_83", "sentence1": "Finally, research into multimodal or multi-view deep learning (Ngiam et al., 2011; Li et al., 2018) offers insights to effectively combine multiple data modalities or views on the same learning problem.", "sentence2": "most work does not directly apply to our problem: i) the audio-text modality is significantly under-represented, ii) the models are typically not required to work online, and iii) most tasks are cast as document-level classification and not sequence labeling (Zadeh et al., 2018).", "label": "contrasting"} {"id": "test_84", "sentence1": "Current ASR approaches rely solely on utilizing audio input to produce transcriptions.", "sentence2": "the wide availability of cameras in smartphones and home devices acts as motivation to build AV-ASR models that rely on and benefit from multimodal input.", "label": "contrasting"} {"id": "test_85", "sentence1": "Even for datasets with dialogue taking place in a similar domain as improv, they naturally contain only a small proportion of yes-ands.", "sentence2": "the relatively large sizes of these datasets still make them useful for dialogue systems.", "label": "contrasting"} {"id": "test_86", "sentence1": "Their model uses an exemplar sentence as a syntactic guide during generation; the generated paraphrase is trained to incorporate the semantics of the input sentence while emulating the syntactic structure of the exemplar (see Appendix D for examples).", "sentence2": "their proposed approach depends on the availability of such exemplars at test time; they manually constructed these for their test set (800 examples).", "label": "contrasting"} {"id": "test_87", "sentence1": "Recent work on controlled generation aims at controlling attributes such as sentiment (Shen et al., 2017), gender or political slant (Prabhumoye et al., 2018), topic (Wang et al., 2017), etc.", "sentence2": "these methods cannot achieve fine-grained control over a property like syntax.", "label": "contrasting"} {"id": "test_88", "sentence1": "In the context of translation, there is often a canonical reordering that should be applied to align better with the target language; for instance, head-final languages like Japanese exhibit highly regular syntax-governed reorderings compared to English.", "sentence2": "in diverse paraphrase generation, there doesn't exist a single canonical reordering, making our problem quite different.", "label": "contrasting"} {"id": "test_89", "sentence1": "Still, dialogue research papers tend to report scores based on word-overlap metrics from the machine translation literature (e.g. BLEU (Papineni et al., 2002), METEOR (Denkowski and Lavie, 2014)).", "sentence2": "word-overlap metrics aggressively penalize the generated response based on lexical differences with the ground truth and correlate poorly to human judgements (Liu et al., 2016).", "label": "contrasting"} {"id": "test_90", "sentence1": "Lowe et al. (2017) propose a learned referenced metric named ADEM, which learns an alignment score be\u0002tween context and response to predict human score annotations.", "sentence2": "since the score is trained to mimic human judgements, it requires collecting large-scale human annotations on the dataset in question and cannot be easily applicable to new datasets (Lowe, 2019).", "label": "contrasting"} {"id": "test_91", "sentence1": "Recently, Tao et al. (2017) proposed a hybrid referenced-unreferenced metric named RUBER, where the metric is trained without requiring hu\u0002man responses by bootstrapping negative samples directly from the dataset.", "sentence2": "referenced metrics (including RUBER, as it is part referenced) are not feasible for evaluation of dialogue models in an online setting-when the model is pitched against a human agent (model-human) or a model agent (model-model)-due to lack of a reference response.", "label": "contrasting"} {"id": "test_92", "sentence1": "All models achieve high scores on the semantic positive samples when only trained with syntactical adversaries.", "sentence2": "training only with syntactical negative samples results in adverse effect on detecting semantic negative items.", "label": "contrasting"} {"id": "test_93", "sentence1": "It has shown that response timings vary based on the semantic content of dialogue responses and the preceding turn (Levinson and Torreira, 2015), and that listeners are sensitive to these fluctuations in timing (Bogels and Levinson, 2017).", "sentence2": "the question of whether certain response timings within different contexts are considered more realistic than others has not been fully investigated.", "label": "contrasting"} {"id": "test_94", "sentence1": "While Step-By-Step uses heuristic string matching to extract plans from the referenced sentences, other methods (GRU and transformer), as well as ours, use plans provided in the enriched WebNLG dataset (Castro Ferreira et al., 2018).", "sentence2": "step-By-step reported worse BLEU results on these plans.", "label": "contrasting"} {"id": "test_95", "sentence1": "GCN does not perform well on Coverage, which demonstrates that the structural gap between encoding and decoding indeed makes generation more difficult.", "sentence2": "it has the smallest difference between Coverage and Faithfulness among all the baselines, indicating that the fidelity of generation can benefit from the encoding of graph-level structural information.", "label": "contrasting"} {"id": "test_96", "sentence1": "These existing neural models have achieved encouraging results.", "sentence2": "when a new condition is added (e.g., a new topic for categorical generation), they require a full retraining or finetuning.", "label": "contrasting"} {"id": "test_97", "sentence1": "Many great works have attempted to solve various subtasks like dialogue generation (Li et al., 2016), poetry generation (Yi et al., 2018) and story generation (Fan et al., 2018) and new techniques keep emerging (Bowman et al., 2016;Yu et al., 2017;Zhou et al., 2020).", "sentence2": "due to the blackbox nature of neural networks, the recent proposed generic models suffer the problem of lacking interpretability and controllability.", "label": "contrasting"} {"id": "test_98", "sentence1": "Previous methods (Kingma et al., 2014; Hu et al., 2017) learn the joint conditional space by jointly considering all conditions.", "sentence2": "once the model is trained, it is not possible to add a new condition without a full retraining.", "label": "contrasting"} {"id": "test_99", "sentence1": "These methods collect user feedback after the model-predicting stage and treat user feedback as additional offline training data to improve the model.", "sentence2": "our model leverages user interaction to increase prediction performance.", "label": "contrasting"} {"id": "test_100", "sentence1": "Pretrained autoregressive models such as GPT (Radford et al., 2018, 2019) are especially capable of generating fluent and coherent text that highly resembles human-written text", "sentence2": "unidirectional attention brings two limitations.", "label": "contrasting"} {"id": "test_101", "sentence1": "This can result in different computational complexity.", "sentence2": "since a typical Graphics Processing Unit (GPU) computes matrices in parallel, the actual difference in inference time is not that significant.", "label": "contrasting"} {"id": "test_102", "sentence1": "The Wikitext103 dataset is more similar to the pretraining datasets, containing long articles.", "sentence2": "the One-Billion Words dataset contains only single sentences, roughly half of which contain less than 24 tokens.", "label": "contrasting"} {"id": "test_103", "sentence1": "Besides, the results show that there are few differences between relative positional embedding and absolute positional embedding for u-PMLM.", "sentence2": "although BERT supports generation in arbitrary word order as well, the PPL for BERT is significantly worse than our proposed u-PMLM for both \"sequential\" and \"random\" settings, demonstrating the effectiveness of the proposed probabilistic masking scheme.", "label": "contrasting"} {"id": "test_104", "sentence1": "We show more cases of text generation in random order for u-PMLM-A and BERT in Appendix B.", "sentence2": "for PPL on One-Billion Words, the performances of u-PMLM and BERT are not satisfactory in comparison with GPT.", "label": "contrasting"} {"id": "test_105", "sentence1": "For GPT, the input text can only be placed in the beginning and the generation process become uncontrollable, resulting in generating sentences with topic drift.", "sentence2": "u-PMLM allows manually placing anchor sentences in the middle or end of the generated text to guide the topic of the generated text.", "label": "contrasting"} {"id": "test_106", "sentence1": "Existing uses of pretrained MLMs in sequenceto-sequence models for automatic speech recognition (ASR) or neural machine translation (NMT) involve integrating their weights (Clinchant et al., 2019) or representations (Zhu et al., 2020) into the encoder and/or decoder during training.", "sentence2": "we train a sequence model independently, then rescore its n-best outputs with an existing MLM.", "label": "contrasting"} {"id": "test_107", "sentence1": "As the MLM gets stronger, the improvement from adding scores from GPT-2 goes to zero, suggesting that their roles overlap at the limit.", "sentence2": "unlike recent work (Shin et al., 2019) but like previous work (Chen et al., 2017), we found that interpolating with a unidirectional LM remained optimal, though our models are trained on different datasets and may have an ensembling effect.", "label": "contrasting"} {"id": "test_108", "sentence1": "In the IID setting, large pretrained Transformer models can attain near human-level performance on numerous tasks (Wang et al., 2019)", "sentence2": "high IID accuracy does not necessarily translate to OOD robustness for image classifiers (Hendrycks and Dietterich, 2019), and pretrained Transformers may embody this same fragility.", "label": "contrasting"} {"id": "test_109", "sentence1": "The recent work of Pruthi et al. (2019), which uses a typo-corrector to defend against adver\u0002sarial typos, is such a reusable defense: it is trained once, then reused across different tasks.", "sentence2": "we find that current typo-correctors do not perform well against even heuristic attacks, limiting their applicability.", "label": "contrasting"} {"id": "test_110", "sentence1": "During training, each occurrence of \"at\" and \"abet\" is replaced with z.", "sentence2": "since \"at\" is much more frequent, classifiers treat z similarly to \"at in order to achieve good overall performance.", "label": "contrasting"} {"id": "test_111", "sentence1": "This data consists of approximately 2 million instances constructed using the abstract and body structure of Wikipedia.", "sentence2": "our ap-proach to pre-training can generate data in unlimited quantity from any text source without assuming a particular document structure.", "label": "contrasting"} {"id": "test_112", "sentence1": "Both improve with multiple iterations, though the improvement is much larger with CMLM.", "sentence2": "even with 10 iterations, ENGINE is comparable to CMLM on DE-EN and outperforms it on RO-EN", "label": "contrasting"} {"id": "test_113", "sentence1": "As in EWISE, in EWISER logits are computed by a dot product between a matrix of hidden scores and output synset embeddings.", "sentence2": "we do not train our own synset embeddings: rather, we employ off-the-shelf vectors.", "label": "contrasting"} {"id": "test_114", "sentence1": "Consequently, the general-language and domain-specific contexts are maximally similar in these cases.", "sentence2": "we assume that the contexts will vary more strongly for basic terms, and for non-terms we do not expect to find domain-specific sentences in the generallanguage corpus at all.", "label": "contrasting"} {"id": "test_115", "sentence1": "Recent research on fairness has primarily focused on racial and gender biases within distributed word representations (Bolukbasi et al., 2016), coreference resolution (Rudinger et al., 2018), sentence encoders (May et al., 2019), and language models .", "sentence2": "we posit that there exists a significant potential for linguistic bias that has yet to be investigated, which is the motivation for our work.", "label": "contrasting"} {"id": "test_116", "sentence1": "In the context of question answering, SpanBERT appears to be slightly more robust than vanilla BERT when comparing overall performance on the two SQuAD datasets.", "sentence2": "the difference becomes significant if we look only at the SQuAD 2.0-fine-tuned models' performance on answerable questions (7% difference).", "label": "contrasting"} {"id": "test_117", "sentence1": "Existing adversarial training approaches have shown that retraining the model on the augmented training set improves robustness (Belinkov and Bisk, 2018; Eger et al., 2019; Jin et al., 2019).", "sentence2": "this requires substantial compute resources.", "label": "contrasting"} {"id": "test_118", "sentence1": "As depicted in Figure 1, we are interested in identifying clusters of subtly distinctive glyph shapes as these correspond to distinct metal stamps in the type-cases used by printers.", "sentence2": "other sources of variation (inking, for example, as depicted in Figure 1) are likely to dominate conventional clustering methods.", "label": "contrasting"} {"id": "test_119", "sentence1": "For example, Gulcehre et al. (2014) explored the influence of selecting different pool\u0002ing norms on the performance of different im\u0002age classification tasks.", "sentence2": "the norms in their method are manually tuned, which are usually very time-consuming and may not be optimal.", "label": "contrasting"} {"id": "test_120", "sentence1": "Multi-task learning (MTL) and transfer learning (TL) are techniques to overcome the issue of data scarcity when training state-of-theart neural networks.", "sentence2": "finding beneficial auxiliary datasets for MTL or TL is a time-and resource-consuming trial-and-error approach.", "label": "contrasting"} {"id": "test_121", "sentence1": "This is reasonable as the \"O\" labels by far make up the majority of all labels in NER datasets.", "sentence2": "this does not help to find similar dataset in other cases, because there is no meaningful ordering of the entropy values when comparing any of the POS samples with all the other samples.", "label": "contrasting"} {"id": "test_122", "sentence1": "The incorporation of pseudo-tags is a standard technique widely used in the NLP community, (Rico et al., 2016; Melvin et al., 2017).", "sentence2": "to the best of our knowledge, our approach is the first attempt to incorporate pseudo-tags as an identification marker of virtual models within a single model.", "label": "contrasting"} {"id": "test_123", "sentence1": "As displayed in Table 2, SINGLEENS surpassed SINGLE by 0.44 and 0.14 on CoNLL-2003 and CoNLL-2000, respectively, for TFM:ELMO with the same parameter size.", "sentence2": "nORMALEnS produced the best results in this setting.", "label": "contrasting"} {"id": "test_124", "sentence1": "Multi-modal neural machine translation (NMT) aims to translate source sentences into a target language paired with images.", "sentence2": "dominant multi-modal NMT models do not fully exploit fine-grained semantic correspondences between semantic units of different modalities, which have potential to refine multi-modal representation learning.", "label": "contrasting"} {"id": "test_125", "sentence1": "Non-autoregressive neural machine translation (NAT) predicts the entire target sequence simultaneously and significantly accelerates inference process.", "sentence2": "nAT discards the dependency information in a sentence, and thus inevitably suffers from the multi-modality problem: the target tokens may be provided by different possible translations, often causing token repetitions or missing.", "label": "contrasting"} {"id": "test_126", "sentence1": "Therefore, the decoder has richer target-side information to detect and recover from such errors.", "sentence2": "it is non-trivial to train the model to learn such behaviour while maintaining a reasonable speedup.", "label": "contrasting"} {"id": "test_127", "sentence1": "The study of calibration on classification tasks has a long history, from statistical machine learning (Platt et al., 1999;Niculescu-Mizil and Caruana, 2005) to deep learning (Guo et al., 2017).", "sentence2": "calibration on structured generation tasks such as neural machine translation (NMT) has not been well studied.", "label": "contrasting"} {"id": "test_128", "sentence1": "Modern neural networks have been found to be miscalibrated on classification tasks in the direction of overestimation (Guo et al., 2017).", "sentence2": "nMT models also suffer from under-estimation problems.", "label": "contrasting"} {"id": "test_129", "sentence1": "For instance, if the spam candidate mutates at the critical positions, the label of the augmented text is likely to change.", "sentence2": "normal candidates are less likely to be affected by this situation.", "label": "contrasting"} {"id": "test_130", "sentence1": "Since manual coding is very laborious and prone to errors, many methods have been proposed for the automatic ICD coding task.", "sentence2": "most of existing methods independently predict each code, ignoring two important characteristics: Code Hierarchy and Code Co-occurrence.", "label": "contrasting"} {"id": "test_131", "sentence1": "(2) The effectiveness of hyperbolic representations: Our proposed model and the CNN+Attention can both correctly predict the code \"518.81\".", "sentence2": "the CNN+Attention model gives contradictory predictions.", "label": "contrasting"} {"id": "test_132", "sentence1": "For example, label \"movie\" should have the largest scalar projection onto a document about \"movie\".", "sentence2": "even the learned label representation of \"music\" can be distinguished from \"movie\", it may also have a large scalar projection onto the document.", "label": "contrasting"} {"id": "test_133", "sentence1": "(Zhao et al., 2019) improves the scalability of capsule networks for MLC.", "sentence2": "they only use CNN to construct capsules, which capture local contextual information (Wang et al., 2016).", "label": "contrasting"} {"id": "test_134", "sentence1": "The typical MLC method SLEEC takes advantage of label correlations by embedding the label co-occurrence graph.", "sentence2": "sLEEC uses TF-IDF vectors to represent documents, thus word order is also ignored.", "label": "contrasting"} {"id": "test_135", "sentence1": "REGGNN is generally superior to both of them as it combines the local and global contextual information dynamically and takes label correlations into consideration using a regularized loss.", "sentence2": "the two capsulebased methods NLP-CAP and HYPERCAPS consistently outperform all the other methods owing to dynamic routing, which aggregates the fine-grained capsule features in a label-aware manner.", "label": "contrasting"} {"id": "test_136", "sentence1": "In the hospitals, the doctors will make a comprehensive analysis mainly based on CC, HPI, PE, TR and the basic information, and make a diagnosis.", "sentence2": "it is very hard for computers to automatically understand all the diverse sections and capture the key information before making an appropriate diagnosis.", "label": "contrasting"} {"id": "test_137", "sentence1": "Zhang et al. (2017) combines the variational auto-encoder and the variational recurrent neural network together to make diagnosis based on laboratory test data.", "sentence2": "laboratory test data are not the only resources considered in this paper.", "label": "contrasting"} {"id": "test_138", "sentence1": "Although ECNN also outputs a probability distribution over all diseases, the result is not interpretable due to its end-to-end nature.", "sentence2": "the interpretability is very important in the CDS to explain how the diagnosis is generated by machines.", "label": "contrasting"} {"id": "test_139", "sentence1": "Remarkable success has been achieved when sufficient labeled training data is available.", "sentence2": "annotating sufficient data is labor-intensive, which establishes significant barriers for generalizing the stance classifier to the data with new targets.", "label": "contrasting"} {"id": "test_140", "sentence1": "The aspect-opinion pairs can provide a global profile about a product or service for consumers and opinion mining systems.", "sentence2": "traditional methods can not directly output aspect-opinion pairs without given aspect terms or opinion terms.", "label": "contrasting"} {"id": "test_141", "sentence1": "As the example sentence shown in Figure 1, (service, great), (prices, great) and (atmosphere, nice friendly) are three aspect-opinion pairs.", "sentence2": "the co-extraction methods can only output the AT set {service, prices, atmosphere} and the OT set {great, nice friendly} jointly.", "label": "contrasting"} {"id": "test_142", "sentence1": "Most of the previous AT and OT extraction meth\u0002ods formulate the task as a sequence tagging prob\u0002lem (Wang et al., 2016, 2017; Wang and Pan, 2018; Yu et al., 2019), specifically using a 5-class tag set: {BA (beginning of aspect), IA (inside of aspect), BP (beginning of opinion), IP (inside of opinion), O (others)}.", "sentence2": "the sequence tagging methods suffer from a huge search space due to the compositionality of labels for extractive ABSA tasks, which has been proven in (Lee et al., 2017b;Hu et al., 2019).", "label": "contrasting"} {"id": "test_143", "sentence1": "Motivated by the correlations between the two tasks, SRL has been utilized to help the ORL task by many previous studies (Ruppenhofer et al., 2008; Marasovic and Frank, 2018; Zhang et al., 2019b).", "sentence2": "when opinion expressions and ar\u0002guments compose complicated syntactic structures, it is difficult to correctly recognize the opinion ar\u0002guments even with shallow semantic representation like SRL (Marasovic and Frank, 2018).", "label": "contrasting"} {"id": "test_144", "sentence1": "Specifically, the pipeline way first trains the dependency parser and then fixes the parser components during training the ORL model.", "sentence2": "the MTL way trains both the parser and the ORL model at the same time.", "label": "contrasting"} {"id": "test_145", "sentence1": "As a baseline, Figure 2-(c) shows the most com\u0002mon MTL method, which shares a common encoder and uses multiple task-specific output layers, known as the hard-parameter-sharing MTL (Ruder, 2017; Marasovic and Frank, 2018).", "sentence2": "this approach is not suitable for our scenario where the auxiliary parsing task has much more labeled data than the main ORL task, since the shared encoder is very likely to bias toward to parsing performance (Xia et al., 2019a).", "label": "contrasting"} {"id": "test_146", "sentence1": "For example, Stab and Gurevych (2017) introduced Argument Annotated Essays (hereafter, Essay), and researchers attempted to predict tree arguments in the corpus (Eger et al., 2017;Potash et al., 2017;Kuribayashi et al., 2019).", "sentence2": "these techniques lack the capability of dealing with more flexible arguments such as reason edges where a proposition can have several parents.", "label": "contrasting"} {"id": "test_147", "sentence1": "Potash et al. (2017) developed a pointer network architec\u0002ture to predict edges.", "sentence2": "we cannot simply utilize them for non-tree arguments because these models were built upon the assumption that an argument forms a tree structure.", "label": "contrasting"} {"id": "test_148", "sentence1": "However, if one wants to apply WSD to some specific corpus, additional annotated training data might be required to meet the similar performance as ours, which defeats the purpose of a weakly supervised setting.", "sentence2": "our contextualization, building upon (Devlin et al., 2019), is adaptive to the input corpus, without requiring any additional human annotations.", "label": "contrasting"} {"id": "test_149", "sentence1": "In micro average, all the span predictions are aggregated together and then compared with the gold spans to get the precision and recall.", "sentence2": "macro average is obtained by calculating the F1 score for each individual sentence and then take an average over all the sentences.", "label": "contrasting"} {"id": "test_150", "sentence1": "Multi-threading is employed since sentences are mutually independent.", "sentence2": "we find that using more than 4 threads does not further improve the speed.", "label": "contrasting"} {"id": "test_151", "sentence1": "We can see that the performance gap is quite steady when we gradually reduce the number of training sentences.", "sentence2": "the gap clearly becomes larger when each training sentence has less annotated dependencies.", "label": "contrasting"} {"id": "test_152", "sentence1": "In BiLSTM-CRF, the CRF layer models the relation between neighbouring labels which leads to better results than simply predicting each label separately based on the BiLSTM outputs.", "sentence2": "the CRF structure models the label sequence globally with the correlations between neighboring labels, which increases the difficulty in distilling the knowledge from the teacher models.", "label": "contrasting"} {"id": "test_153", "sentence1": "In particular, discarding the conversion matrix in the ESD module also leads to the performance drop, which indicates the usefulness of capturing the label correspondence between the auxiliary module and our main MNER task.", "sentence2": "as the main contribution of our MMI module, Image-Aware Word Representations (WR) demonstrates its indispensable role in the final performance due to the moderate performance drop after removal.", "label": "contrasting"} {"id": "test_154", "sentence1": "These neural approaches have been shown to achieve the state-of-the-art performance on different benchmark datasets based on formal text (Yang et al., 2018).", "sentence2": "when applying these approaches to social media text, most of them fail to achieve satisfactory results.", "label": "contrasting"} {"id": "test_155", "sentence1": "Despite not being exposed to explicit syntactic supervision, neural language models (LMs), such as recurrent neural networks, are able to generate fluent and natural sentences, suggesting that they induce syntactic knowledge about the language to some extent.", "sentence2": "it is still under debate whether such induced knowledge about grammar is robust enough to deal with syntactically challenging constructions such as long-distance subjectverb agreement.", "label": "contrasting"} {"id": "test_156", "sentence1": "We expect that the main reason for lower performance for object RCs is due to frequency, and with our augmentation the accuracy will reach the same level as that for subject RCs.", "sentence2": "for both all and animate cases, accuracies are below those for subject rCs (Figure 2).", "label": "contrasting"} {"id": "test_157", "sentence1": "Moreover, Huang et al. (2019) im\u0002prove TextGCN by introducing the message pass\u0002ing mechanism and reducing the memory consump\u0002tion", "sentence2": "there are two major drawbacks in these graph-based methods.", "label": "contrasting"} {"id": "test_158", "sentence1": "A concurrent work (Warstadt et al., 2019b) facilitates diagnosing language models by creating linguistic minimal pairs datasets for 67 isolate grammatical paradigms in English using linguistcrafted templates.", "sentence2": "we do not rely heavily on artificial vocabulary and templates.", "label": "contrasting"} {"id": "test_159", "sentence1": "Such attention weights measure the relative importance of the token within a specific input sequence.", "sentence2": "the attention score a j captures the absolute importance of the token.", "label": "contrasting"} {"id": "test_160", "sentence1": "After that, both of them directly use the word representation of two languages to retrieve the initial bilingual lexicons by computing the cosine distances of source and target word representations.", "sentence2": "directly finding word alignments from scratch has some demerits.", "label": "contrasting"} {"id": "test_161", "sentence1": "Unsupervised neural machine translation (UNMT) has recently achieved remarkable results for several language pairs.", "sentence2": "it can only translate between a single language pair and cannot produce translation results for multiple language pairs at the same time.", "label": "contrasting"} {"id": "test_162", "sentence1": "For example, Xu et al. (2019) and Sen et al. (2019) proposed a multilingual scheme that jointly trains multiple languages with multiple decoders.", "sentence2": "the performance of their MUNMT is much worse than our re-implemented individual baselines (shown in Tables 2 and 3) and the scale of their study is modest (i.e., 4-5 languages).", "label": "contrasting"} {"id": "test_163", "sentence1": "Moreover, the MUNMT model could alleviate the poor performance achieved with low-resource language pairs, such as En-Lt and En-Lv.", "sentence2": "the performance of MUNMT is slightly worse than SM in some language pairs.", "label": "contrasting"} {"id": "test_164", "sentence1": "The standard training algorithm in neural machine translation (NMT) suffers from exposure bias, and alternative algorithms have been proposed to mitigate this.", "sentence2": "the practical impact of exposure bias is under debate.", "label": "contrasting"} {"id": "test_165", "sentence1": "Previous work has sought to reduce exposure bias in training (Bengio et al., 2015; Ranzato et al., 2016; Shen et al., 2016; Wiseman and Rush, 2016; Zhang et al., 2019).", "sentence2": "the relevance of error propagation is under debate: Wu et al. (2018) argue that its role is overstated in literature, and that linguistic features explain some of the accuracy drop at higher time steps.", "label": "contrasting"} {"id": "test_166", "sentence1": "Advanced pre-trained models for text representation have achieved state-of-the-art performance on various text classification tasks.", "sentence2": "the discrepancy between the semantic similarity of texts and labelling standards affects classifiers, i.e. leading to lower performance in cases where classifiers should assign different labels to semantically similar texts.", "label": "contrasting"} {"id": "test_167", "sentence1": "In general, AAN achieved greater performance than AM.", "sentence2": "their effectiveness turned out to be task-dependent.", "label": "contrasting"} {"id": "test_168", "sentence1": "Previous studies aimed to improve multiple tasks; hence, they required multiple sets of annotated datasets.", "sentence2": "our method does not require any extra labelled datasets and is easily applicable to various classification tasks.", "label": "contrasting"} {"id": "test_169", "sentence1": "On top of it, Crosslingual training (or bilingual, denoted by \"Cross\") obtains marginal improvements for moderately lowresource languages.", "sentence2": "the performance drops dramatically for two extremely low-resource languages, i.e., JA from 0.740 to 0.711 and EL from 0.702 to 0.684.", "label": "contrasting"} {"id": "test_170", "sentence1": "Previous studies in multimodal sentiment analysis have used limited datasets, which only contain unified multimodal annotations.", "sentence2": "the unified annotations do not always reflect the independent sentiment of single modalities and limit the model to capture the difference between modalities.", "label": "contrasting"} {"id": "test_171", "sentence1": "An intuitive idea is that the greater the difference between inter-modal representations, the better the complementarity of intermodal fusion.", "sentence2": "it is not easy for existing late-fusion models to learn the differences between different modalities, further limits the performance of fusion.", "label": "contrasting"} {"id": "test_172", "sentence1": "The CHEAVD (Li et al., 2017) is also a Chinese multimodal dataset, but it only contains two modalities (vision and audio) and one unified annotation.", "sentence2": "sIMs has three modalities and unimodal annotations except for multimodal annotations for each clip.", "label": "contrasting"} {"id": "test_173", "sentence1": "However, these existing multimodal datasets only contain a unified multimodal annotation for each multimodal corpus.", "sentence2": "sIMs contains both unimodal and multimodal annotations.", "label": "contrasting"} {"id": "test_174", "sentence1": "Hierarchical FM performs better than MLP+CNN by incorporating additional attributes that provide the visual semantic information and generating better feature representations via a hierarchical fusion framework.", "sentence2": "these multimodal baselines pay more attention to the fusion of multimodal features.", "label": "contrasting"} {"id": "test_175", "sentence1": "Thus, their performances are worse than MIARN, which focuses on textual context to model the contrast information between individual words and phrases.", "sentence2": "due to the nature of short text, relying on textual information is often insufficient, especially in multimodal tweets where cross-modality context relies the most important role.", "label": "contrasting"} {"id": "test_176", "sentence1": "As interpretability is important for understanding and debugging the translation process and particularly to further improve NMT models, many efforts have been devoted to explanation methods for NMT (Ding et al., 2017;Alvarez-Melis and Jaakkola, 2017;Li et al., 2019;Ding et al., 2019;.", "sentence2": "little progress has been made on evaluation metric to study how good these explanation methods are and which method is better than others for NMT.", "label": "contrasting"} {"id": "test_177", "sentence1": "In terms of i), Word Alignment Error Rate (AER) can be used as a metric to evaluate an explanation method by measuring agreement between human-annotated word alignment and that derived from the explanation method.", "sentence2": "aER can not measure explanation methods on those target words that are not aligned to any source words according to human annotation.", "label": "contrasting"} {"id": "test_178", "sentence1": "On one hand, the real data distribution of c t is unknowable, making it impossible to exactly define the expectation with respect to an unknown distribution.", "sentence2": "the domain of a proxy model Q is not bounded, and it is difficult to minimize a model Q within an unbounded domain.", "label": "contrasting"} {"id": "test_179", "sentence1": "This backpropagation through the generated data, combined with adversarial learning instabilities, has proven to be a compelling challenge when applying GANs for discrete data such as text.", "sentence2": "it remains unknown if this is also an issue for feature matching networks since the effectiveness of GFMN for sequential discrete data has not yet been studied.", "label": "contrasting"} {"id": "test_180", "sentence1": "An interesting comparison would be between Se-qGFMN and GANs that use BERT as a pre-trained discriminator.", "sentence2": "gANs fail to train when a very deep network is used as the discriminator Moreover, SeqgFMN also outperforms gAN generators even when shallow word embeddings (glove / FastText) are used to perform feature matching.", "label": "contrasting"} {"id": "test_181", "sentence1": "Extractive MRC requires a model to extract an answer span to a question from reference documents, such as the tasks in SQuAD (Rajpurkar et al., 2016) and CoQA (Reddy et al., 2019).", "sentence2": "non-extractive MRC infers answers based on some evidence in reference documents, including Yes/No question answering (Clark et al., 2019), multiple-choice MRC (Lai et al., 2017; Khashabi et al., 2018; Sun et al., 2019), and open domain question answering (Dhingra et al., 2017b).", "label": "contrasting"} {"id": "test_182", "sentence1": "RL methods can indeed train a better extractor without evidence labels.", "sentence2": "they are much more complicated and unstable to train, and highly dependent on model pre-training.", "label": "contrasting"} {"id": "test_183", "sentence1": "As the innermost ring shows, about 80% of the evidence predicted by BERT-HA (iter 0) was incorrect.", "sentence2": "the proportion of wrong instances reduced to 60% after self-training (iter 3).", "label": "contrasting"} {"id": "test_184", "sentence1": "Earlier studies have attempted to perform the MWP task via statistical machine learning methods (Kushman et al., 2014; Hosseini et al., 2014; Mitra and Baral, 2016; Roy and Roth, 2018) and seman\u0002tic parsing approaches (Shi et al., 2015; Koncel\u0002Kedziorski et al., 2015; Roy and Roth, 2015; Huang et al., 2017).", "sentence2": "these methods are nonscalable as tremendous efforts are required to design suitable features and expression templates.", "label": "contrasting"} {"id": "test_185", "sentence1": "To enrich the representation of a quantity, the relationships between the descriptive words associated with a quantity need to be modeled.", "sentence2": "such relationships cannot be effectively modeled using recurrent models, which are commonly used in the existing MWP deep learning methods.", "label": "contrasting"} {"id": "test_186", "sentence1": "While all of these methods are bag-of-words models, Liu et al. (2019a) recently proposed an architecture based on context2vec (Melamud et al., 2016).", "sentence2": "in contrast to our work, they (i) do not incorporate surface-form information and (ii) do not directly access the hidden states of context2vec, but instead simply use its output distribution.", "label": "contrasting"} {"id": "test_187", "sentence1": "For this reason, previous works used an in-house mapping between BabelNet versions to make them up to date.", "sentence2": "in this process, several gold instances were lost making the datasets smaller than the original ones.", "label": "contrasting"} {"id": "test_188", "sentence1": "Word-Net provides information about sense frequency that is either manually-annotated or derived from SemCor (Miller et al., 1993), i.e., a corpus where words are manually tagged with WordNet meanings.", "sentence2": "neither WordNet nor SemCor have been updated in the past 10 years, thus making their information about sense frequency outdated.", "label": "contrasting"} {"id": "test_189", "sentence1": "The method of applying REINFORCE to the discriminative parser is straightforward because sampling trees from the discriminative parser is easy.", "sentence2": "that is not the case for the generative model from which we have to sample both trees and sentences at the same time.", "label": "contrasting"} {"id": "test_190", "sentence1": "In this task, models are expected to make predictions with the semantic information rather than with the demographic group identity information (e.g., \"gay\", \"black\") contained in the sentences.", "sentence2": "recent research points out that there widely exist some unintended biases in text classification datasets.", "label": "contrasting"} {"id": "test_191", "sentence1": "In other words, a non-discrimination model should perform similarly across sentences containing different demographic groups.", "sentence2": "\"perform similarly\" is indeed hard to define.", "label": "contrasting"} {"id": "test_192", "sentence1": "Similar to results on Toxicity Comments, we find that both Weight and Supplement perform significantly better than Baseline in terms of IPTTS AUC and FPED, and the results of Weight and Supplement are comparable.", "sentence2": "we notice that Weight and Supplement improve FNED slightly, while the differences are not statistically significant at confidence level 0.05.", "label": "contrasting"} {"id": "test_193", "sentence1": "Current approaches define interpretation in a rather ad-hoc manner, motivated by practical usecases and applications.", "sentence2": "this view often fails to distinguish between distinct aspects of the interpretation's quality, such as readability, plausibility and faithfulness (Herman, 2017).", "label": "contrasting"} {"id": "test_194", "sentence1": "For example, Serrano and Smith (2019) and Jain and Wallace (2019) show that high attention weights need not necessarily correspond to a higher impact on the model's predictions and hence they do not provide a faithful explanation for the model's predictions.", "sentence2": "wiegreffe and Pinter (2019) argues that there is still a possibility that attention distributions may provide a plausible explanation for the predictions.", "label": "contrasting"} {"id": "test_195", "sentence1": "Our main goal is to show that our proposed models provide more faithful and plausible explanations for their predictions.", "sentence2": "before we go there we need to show that the predictive performance of our models is comparable to that of a vanilla LSTM model and significantly better than non-contextual models.", "label": "contrasting"} {"id": "test_196", "sentence1": "We observe that randomly permuting the attention weights in the Diversity and Orthogonal LSTM model results in significantly different outputs.", "sentence2": "there is little change in the vanilla LSTM model's output for several datasets suggesting that the attention weights are not so meaningful.", "label": "contrasting"} {"id": "test_197", "sentence1": "Several other works (Shao et al., 2019; Martins and Astudillo, 2016; Malaviya et al., 2018; Niculae and Blondel, 2017; Maruf et al., 2019; Peters et al., 2018) focus on improving the interpretability of attention distributions by inducing sparsity.", "sentence2": "the extent to which sparse attention distributions actually offer faithful and plausible explanations haven't been studied in detail.", "label": "contrasting"} {"id": "test_198", "sentence1": "The plotted eight words are gathered together, and it can be seen that hidden representations of the same word gather in the same place regardless of correctness.", "sentence2": "fine-tuned BERT produces a vector space that demonstrates correct and incorrect words on different sides, showing that hidden representations take grammatical errors into account when fine-tuned on GEC corpora.", "label": "contrasting"} {"id": "test_199", "sentence1": "For instance, DKN (Wang et al., 2018) learns knowledge-aware news representation via multi-channel CNN and gets a representation of a user by aggregating her clicked news history with different weights.", "sentence2": "these methods (Wu et al., 2019b; Zhu et al., 2019; An et al., 2019) usu\u0002ally focus on news contents, and seldom consider the collaborative signal in the form of high-order connectivity underlying the user-news interactions.", "label": "contrasting"} {"id": "test_200", "sentence1": "Wang et al. (2019) explored the GNN to capture high-order connectivity information in user-item graph by propagating embeddings on it, which achieves better performance on recommendation.", "sentence2": "existing news recommendation methods focus on, and rely heavily on news contents.", "label": "contrasting"} {"id": "test_201", "sentence1": "Initially, we set z u,k = s u,k .", "sentence2": "after obtaining the latent variables {r d,k }, we can find an estimate of z u,k by aggregating information from the clicked news, which is computed as Eq.", "label": "contrasting"} {"id": "test_202", "sentence1": "Most existing methods usually learn the representations of users and news from news contents for recommendation.", "sentence2": "they seldom consider highorder connectivity underlying the user-news interactions.", "label": "contrasting"} {"id": "test_203", "sentence1": "It supposes that comparing to less important roles, the roles with bigger impact are expected to appear at more places and are more evenly distributed over the story.", "sentence2": "this assumption ignores actions of roles (denoted as behavioral semantic information), which may be a key factor that estimates their impacts in legalcontext scenarios.", "label": "contrasting"} {"id": "test_204", "sentence1": "Position or frequency information does not effectively reflect the status of a role in such samples.", "sentence2": "our method captures this information by the cooperation mode feature between Yin and Zhao, with the help of verb \"instructed\".", "label": "contrasting"} {"id": "test_205", "sentence1": "As we can see from Figure 2(a), the plain nets suffer from the degradation problem, which is not caused by overfitting, as they exhibit lower training BLEU.", "sentence2": "the 72-layer MSC exhibits higher training BLEU than the 36-layer counterpart and is generalizable to the validation data.", "label": "contrasting"} {"id": "test_206", "sentence1": "Figure 4A) shows, as first identified by Kozlowski et al. (2019), that much of this is due to the variance of the survey data along that dimension; the correlation between variance and the coeffi\u0002cients in Figure 3 is 0.91", "sentence2": "as discussed above, Kozlowski et al. (2019) study more general concepts on more general dimensions, and note that they have no easy way to connect their observations to any critical social processes. ", "label": "contrasting"} {"id": "test_207", "sentence1": "On the other hand, from the \"bias\" per-spective, this suggests that a vast array of social biases are encoded in embeddings.", "sentence2": "we also find that some beliefs-specifically, extreme beliefs on salient dimensions -are easier to measure than others.", "label": "contrasting"} {"id": "test_208", "sentence1": "Attention mechanisms learn to assign soft weights to (usually contextualized) token representations, and so one can extract highly weighted tokens as rationales.", "sentence2": "attention weights do not in general provide faithful explanations for predictions (Jain and Wallace, 2019; Serrano and Smith, 2019; Wiegreffe and Pinter, 2019; Zhong et al., 2019; Pruthi et al., 2020; Brunner et al., 2020; Moradi et al., 2019; Vashishth et al., 2019).", "label": "contrasting"} {"id": "test_209", "sentence1": "Original rationale annotations were not necessarily comprehensive; we thus collected comprehensive rationales on the final two folds of the original dataset (Pang and Lee, 2004).", "sentence2": "to most other datasets, the rationale annotations here are span level as opposed to sentence level.", "label": "contrasting"} {"id": "test_210", "sentence1": "In general, the rationales we have for tasks are sufficient to make judgments, but not necessarily comprehensive.", "sentence2": "for some datasets we have explicitly collected comprehensive rationales for at least a subset of the test set.", "label": "contrasting"} {"id": "test_211", "sentence1": "Since DeFormer retains much of the original structure, we can initialize this model with the pre-trained weights of the original Transformer and fine-tune directly on downstream tasks.", "sentence2": "deFormer looses some information in the representations of the lower layers.", "label": "contrasting"} {"id": "test_212", "sentence1": "The upper layers can learn to compensate for this during finetuning.", "sentence2": "we can go further and use the original model behavior as an additional source of supervision.", "label": "contrasting"} {"id": "test_213", "sentence1": "This is an orthogonal approach that can be combined with our decomposition idea.", "sentence2": "for the paired-input tasks we consider, pruning heads only provides limited speedup.", "label": "contrasting"} {"id": "test_214", "sentence1": "This is because the candidate justifications are coming from a relatively small numbers of paragraphs in MultiRC; thus even shorter queries (= 2 words) can retrieve relevant justifications.", "sentence2": "the number of candidate justifications in QASC is much higher, which requires longer queries for disambiguation (>= 4 words).", "label": "contrasting"} {"id": "test_215", "sentence1": "Access to such data can greatly facilitate investigation of phonetic typology at a large scale and across many languages.", "sentence2": "it is nontrivial and computationally intensive to obtain such alignments for hundreds of languages, many of which have few to no resources presently available.", "label": "contrasting"} {"id": "test_216", "sentence1": "In addition to its coverage, the CMU Wilderness corpus is unique in two additional aspects: cleanly recorded, read speech exists for all languages in the corpus, and the same content (modulo translation) exists across all languages.", "sentence2": "this massively multilingual speech corpus is challenging to work with directly.", "label": "contrasting"} {"id": "test_217", "sentence1": "Since our greedy opportunistic decoding doesn't change the final output, there is no difference in BLEU compared with normal decoding, but the latency is reduced.", "sentence2": "by applying beam search, we can achieve 3.1 BLEU improvement and 2.4 latency reduction on wait-7 policy.", "label": "contrasting"} {"id": "test_218", "sentence1": "Thanks to the wealth of high-quality annotated images available in popular repositories such as ImageNet, multimodal language-vision research is in full bloom.", "sentence2": "events, feelings and many other kinds of concepts which can be visually grounded are not well represented in current datasets.", "label": "contrasting"} {"id": "test_219", "sentence1": "On one hand, the inclusion of NC concepts would be an important step towards wide-coverage image semantic understanding.", "sentence2": "it also goes in the same direction as recent mul\u0002timodal language-vision approaches, e.g., mono\u0002and cross-lingual Visual Sense Disambiguation (Barnard and Johnson, 2005; Loeff et al., 2006; Saenko and Darrell, 2008; Gella et al., 2016, 2019).", "label": "contrasting"} {"id": "test_220", "sentence1": "Our experiments show that both systems are reliable on our task, achieving precision and F1 scores that are over 70% on all the splits (see Table 2).", "sentence2": "the F-VLP model proves to be the most stable for the task.", "label": "contrasting"} {"id": "test_221", "sentence1": "This accordance to some extent verifies that the neurons found through influence paths are functionally important.", "sentence2": "the t-values shown in Table 1 show that both neuron 125 and 337 are influential regardless of the subject number, whereas Lakretz et al. assign a subject number for each of these two neurons due to their disparate effect in lowering accuracy in ablation experiment.", "label": "contrasting"} {"id": "test_222", "sentence1": "A slightly worse NA task performance (Lakretz et al., 2019) in cases of attractors (SP, PS) indicates that they interfere with prediction of the correct verb.", "sentence2": "we also observe that helpful nouns (SS, PP) contribute positively to the correct verb number (although they should not from a grammar perspective).", "label": "contrasting"} {"id": "test_223", "sentence1": "In particular, in the largest setting with N = 1M, the BERT-24 embeddings distilled from the best-performing layer for each dataset drastically outperform both Word2Vec and GloVe.", "sentence2": "this can be seen as an unfair comparison given that we are selecting specific layers for specific datasets.", "label": "contrasting"} {"id": "test_224", "sentence1": "As such, the validity of treating the resulting static embeddings as reliable proxies for the original contextualized model still remains open.", "sentence2": "human language processing has often been conjectured to have both context-dependent and context-independent properties (Barsalou, 1982; Rubio-Fernandez, 2008; Depraetere, 2014, 2019).", "label": "contrasting"} {"id": "test_225", "sentence1": "We find that each of the tested word features can be encoded in contextual embeddings for other words of the sentence, often with perfect or nearperfect recoverability.", "sentence2": "we see substantial variation across encoders in how robustly each information type is distributed to which tokens.", "label": "contrasting"} {"id": "test_226", "sentence1": "CheckList provides a framework for such techniques to systematically evaluate these alongside a variety of other capabilities.", "sentence2": "checkList cannot be directly used for non-behavioral issues such as data versioning problems (Amershi et al., 2019), labeling errors, annotator biases (Geva et al., 2019), worst-case security issues (Wallace et al., 2019), or lack of interpretability (Ribeiro et al., 2016).", "label": "contrasting"} {"id": "test_227", "sentence1": "These experimental results show the critical role of triggers in dialogue-based relation extraction.", "sentence2": "trigger identification is perhaps as difficult as relation extraction, and it is labor-intensive to annotate large-scale datasets with triggers.", "label": "contrasting"} {"id": "test_228", "sentence1": "RST Graph is constructed from RST parse trees over EDUs of the document.", "sentence2": "coreference Graph connects entities and their coreference clusters/mentions across the document.", "label": "contrasting"} {"id": "test_229", "sentence1": "As observed in Louis et al. (2010), the RST tree structure already serves as a strong indicator for content selection.", "sentence2": "the agreement between rhetorical relations tends to be lower and more ambiguous.", "label": "contrasting"} {"id": "test_230", "sentence1": "BERT is originally trained to encode a single sentence or sentence pair.", "sentence2": "a news article typically contains more than 500 words, hence we need to make some adaptation to apply BERT for document encoding.", "label": "contrasting"} {"id": "test_231", "sentence1": "Because of the similarity to our task, we use a BERT-based neural network as the architecture for the coverage model.", "sentence2": "the coverage task differs from MLM in two ways.", "label": "contrasting"} {"id": "test_232", "sentence1": "Just as with the Summarizer, by using a standardized architecture and model size, we can make use of pretrained models.", "sentence2": "it is important for Fluency to fine tune the language model on the target domain, so that the Summarizer is rewarded for generating text similar to target content.", "label": "contrasting"} {"id": "test_233", "sentence1": "EMONET was conceived as a multiclass classification task for Plutchik-8 emotions (Abdul-Mageed and Ungar, 2017).", "sentence2": "we introduce binary classification tasks, one for each Plutchik-8 emotion.", "label": "contrasting"} {"id": "test_234", "sentence1": "In our work, we raise similar concerns but through a different angle by highlighting issues with the evaluation procedure used by several recent methods. Chandrahas et al. (2018) analyze the geometry of KG embeddings and its correlation with task performance while Nayyeri et al. (2019) examine the effect of different loss functions on performance.", "sentence2": "their analysis is restricted to non-neural approaches.", "label": "contrasting"} {"id": "test_235", "sentence1": "Several recently proposed methods report high performance gains on a particular dataset.", "sentence2": "their performance on another dataset is not consistently improved.", "label": "contrasting"} {"id": "test_236", "sentence1": "Recently many reading comprehension datasets requiring complex and compositional reasoning over text have been introduced, including HotpotQA (Yang et al., 2018), DROP (Dua et al., 2019), Quoref , and ROPES (Lin et al., 2019).", "sentence2": "models trained on these datasets (Hu et al., 2019;Andor et al., 2019) only have the final answer as supervision, leaving the model guessing at the correct latent reasoning.", "label": "contrasting"} {"id": "test_237", "sentence1": "Recent proposed approaches have made promising progress in dialogue state tracking (DST).", "sentence2": "in multi-domain scenarios, ellipsis and reference are frequently adopted by users to express values that have been mentioned by slots from other domains.", "label": "contrasting"} {"id": "test_238", "sentence1": "Open vocabulary models show the promising performance in multidomain DST.", "sentence2": "ellipsis and reference phenomena among multi-domain slots are still less explored in existing literature.", "label": "contrasting"} {"id": "test_239", "sentence1": "Some proposed solutions rely on leveraging knowledge distillation in the pre-training step, e.g., (Sanh et al., 2019), or used parameter reduction techniques (Lan et al., 2019) to reduce inference cost.", "sentence2": "the effectiveness of these approaches varies depending on the target task they have been applied to.", "label": "contrasting"} {"id": "test_240", "sentence1": "This may be seen as a generalization of the ST approach, where the student needs to learn a simpler task than the teacher.", "sentence2": "our approach is significantly different from the traditional ST setting, which our preliminary investigation showed to be not very effective.", "label": "contrasting"} {"id": "test_241", "sentence1": "These two dialogue-specific LM approaches, ULM and UOP, give very marginal improvement over the baseline models, that is rather surprising.", "sentence2": "they show good improvement when combined with UID, implying that pre-training language models may not be enough to enhance the performance by itself but can be effective when it is coupled with an appropriate fine-tuning approach.", "label": "contrasting"} {"id": "test_242", "sentence1": "Recently, pretrained language representation models (Kocijan et al., 2019;Radford et al., 2019;Liu et al., 2019) have demonstrated significant improvements in both unsupervised and supervised settings.", "sentence2": "as these approaches treat the concept 'commonsense knowledge' as a black box, we are not clear about why they can do better (e.g., can these models understand commonsense or they just capture the statistical bias of the dataset) and do not know how to further improve them.", "label": "contrasting"} {"id": "test_243", "sentence1": "For evaluation purposes, we may have labeled documents in the target language.", "sentence2": "they are only used during the test period.", "label": "contrasting"} {"id": "test_244", "sentence1": "For the last layer before softmax, even though XLM-FT also generates reasonable representations to separate positive and negative reviews, the data points are scattered randomly.", "sentence2": "our model's output in the lower right panel of Figure 3 shows two more obvious clusters with corresponding labels that can be easily separated.", "label": "contrasting"} {"id": "test_245", "sentence1": "This rating method, also known as Likert scale or Mean Opinion Score, is known to have two major drawbacks (Ye and Doermann, 2013): (1) Absolute rating is often treated as if it produces data on an interval scale.", "sentence2": "assessors rarely perceive labels as equidistant, thus producing only ordinal data.", "label": "contrasting"} {"id": "test_246", "sentence1": "Overall, the Bradley-Terry model appears to be a promising candidate for our purposes: its robustness and statistical properties have been studied in great detail (Hunter, 2004), and it can be efficiently computed (Chen et al., 2013).", "sentence2": "an alternative offline sampling method has to be formulated, which we introduce in the following section.", "label": "contrasting"} {"id": "test_247", "sentence1": "This can be generalized to higher step sizes s: for instance, if s = 2, all items that are separated by two positions around the ring are compared.", "sentence2": "this strategy suffers from the major drawback that for some step sizes, the resulting graph has multiple unconnected components, thus violating the restriction that the comparison matrix must form a strongly connected graph.", "label": "contrasting"} {"id": "test_248", "sentence1": "Using a higher temperature yields a softer attention distribution.", "sentence2": "a sharper attention distribution might be more suitable for NER because only a few tokens in the sentence are named entities.", "label": "contrasting"} {"id": "test_249", "sentence1": " In this work, we follow Chen et al. (2019) and use exactly the same functions.", "sentence2": "as shown in 7 (c), understanding this statement requires the function of difference time, which is not covered by the current set.", "label": "contrasting"} {"id": "test_250", "sentence1": "Word-level attacking, which can be regarded as a combinatorial optimization problem, is a well-studied class of textual attack methods.", "sentence2": "existing word-level attack models are far from perfect, largely because unsuitable search space reduction methods and inefficient optimization algorithms are employed.", "label": "contrasting"} {"id": "test_251", "sentence1": "Nonetheless, its attack validity rates against BiLSTM and BERT on SST-2 dramatically fall to 59.5% and 56.5%.", "sentence2": "ours are 70.5% and 72.0%, and their differences are significant according to the results of significance tests in Appendix D. In this section, we conduct detailed decomposition analyses of different word substitution methods (search space reduction methods) and different search algorithms, aiming to further demonstrate the advantages of our sememe-based word substitution method and PSO-based search algorithm.", "label": "contrasting"} {"id": "test_252", "sentence1": "Mixed counting models that use the negative binomial distribution as the prior can well model over-dispersed and hierarchically dependent random variables; thus they have attracted much attention in mining dispersed document topics.", "sentence2": "the existing parameter inference method like Monte Carlo sampling is quite time-consuming.", "label": "contrasting"} {"id": "test_253", "sentence1": "On the one hand, NVI-based models are fast and easy to estimate but hard to interpret.", "sentence2": "document modeling via mixed counting models is easy to interpret but difficult to infer.", "label": "contrasting"} {"id": "test_254", "sentence1": "Such a framework learns the distribution of input data well, enabling it to combine with the traditional probability graphical models (e.g., LDA) and infer model parameters quickly (Srivastava and Sutton, 2017).", "sentence2": "how to effectively integrate the distributed dependencies in mixed counting models into the framework of variational inference is still quite a challenging problem.", "label": "contrasting"} {"id": "test_255", "sentence1": "From these results, we can observe that the proportion of topics obtained by NB-NTM is close to the topic-word number distribution.", "sentence2": "gNB-NTM obtains more dispersed proportions of topics than NB-NTM.", "label": "contrasting"} {"id": "test_256", "sentence1": "Everything was a point in a vector space, and everything about the nature of language could be learned from data.", "sentence2": "most computational linguists had linguistic theories and the poverty-of-thestimulus argument.", "label": "contrasting"} {"id": "test_257", "sentence1": "And deep learning allows many aspects of these structured representations to be learned from data.", "sentence2": "successful deep learning architectures for natural language currently still have many handcoded aspects.", "label": "contrasting"} {"id": "test_258", "sentence1": "(Alex et al., 2007) propose several different modeling techniques (layering and cascading) to combine multiple CRFs for nested NER.", "sentence2": "their approach cannot handle nested entities of the same entity type.", "label": "contrasting"} {"id": "test_259", "sentence1": "Here, the two central entities (both indicate the city Liverpool) have similar sizes of neighborhoods and three common neighbors.", "sentence2": "the three common neighbors (indicate United Kingdom, England and Labour Party (UK), respectively) are not discriminative enough.", "label": "contrasting"} {"id": "test_260", "sentence1": "Most similar to our work are (Chen et al., 2019) and (Ghosh et al., 2019), as both studies are considered using the concept of question answering to address NLVL.", "sentence2": "both studies do not explain the similarity and differences between NLVL and traditional span-based QA, and they do not adopt the standard span-based QA framework.", "label": "contrasting"} {"id": "test_261", "sentence1": "By treating input video as text passage, the above frameworks are all applicable to NLVL in principle.", "sentence2": "these frameworks are not designed to consider the differences between video and text passage.", "label": "contrasting"} {"id": "test_262", "sentence1": "Figure 9 shows that VSLBase tends to predict longer moments, e.g., more samples with length error larger than 4 seconds in Charades-STA or 30 seconds in Activ-ityNet.", "sentence2": "constrained by QGH, VSLNet tends to predict shorter moments, e.g., more samples with length error smaller that -4 seconds in Charades-STA or -20 seconds in ActivityNet Caption.", "label": "contrasting"} {"id": "test_263", "sentence1": "A very recent work makes use of attention over spans instead of syntactic distance to inject inductive bias to language models (Peng et al., 2019).", "sentence2": "the time complexity of injecting supervision is much higher than distancebased approach (O(n 2 ) VS O(n) ).", "label": "contrasting"} {"id": "test_264", "sentence1": "For example, all punctuation symbols are removed, all characters are lower-cased, the vocabulary size is truncated at 10,000 and all sentences are concatenated.", "sentence2": "this version of PTB discards the parse tree structures, which makes it unsuitable for comparing sequential language models with those utilizing tree structures.", "label": "contrasting"} {"id": "test_265", "sentence1": "Because of the sequential and parallel nature of our model, it can directly inherit and benefit from this set of tricks.", "sentence2": "it is non-trivial to use them for RNNG and URNNG.", "label": "contrasting"} {"id": "test_266", "sentence1": "In particular, Kim et al. (2019a) report that unsupervised URNNG achieves 45.4 WSJ F1 in a similar setting, while another URNNG that finetunes a supervised RNNG model gives a much better F1 of 72.8, leading a 27.4 F1 improvement.", "sentence2": "the F1 of our structure prediction trees is 61.3 in unbiased algorithm.", "label": "contrasting"} {"id": "test_267", "sentence1": "But this strategy does not consider that it is properly difficult to acquire a dictionary with high quality for a brand new domain.", "sentence2": "we develop a simple and efficient strategy to perform domain-specific words mining without any predefined dictionaries.", "label": "contrasting"} {"id": "test_268", "sentence1": "In some languages, predicting declension class is argued to be easier if we know the noun's phonological form (Aronoff, 1992;Dressler and Thornton, 1996) or lexical semantics (Carstairs-McCarthy, 1994;Corbett and Fraser, 2000).", "sentence2": "semantic and phonological clues are, at best, only very imperfect hints as to class (Wurzel, 1989;Harris, , 1992Aronoff, 1992;Halle and Marantz, 1994;Corbett and Fraser, 2000;Aronoff, 2007).", "label": "contrasting"} {"id": "test_269", "sentence1": "Until now, we have assumed a one-to-one mapping between paradigm slots and surface form changes.", "sentence2": "different EDIT TREES may indeed represent the same inflection.", "label": "contrasting"} {"id": "test_270", "sentence1": "Our system, thus, needs to learn to apply the correct transformation for a combination of lemma and paradigm slot.", "sentence2": "mapping lemmas and paradigm slots to inflected forms corresponds exactly to the morphological inflection task, which has been the subject of multiple shared tasks over the last years (Cotterell et al., 2018).", "label": "contrasting"} {"id": "test_271", "sentence1": "Systems for supervised or semi-supervised paradigm completion are commonly being evalu\u0002ated using word-level accuracy (Dreyer and Eisner, 2011; Cotterell et al., 2017).", "sentence2": "this is not possible for our task because our system cannot access the gold data paradigm slot descriptions and, thus, does not necessarily produce one word for each ground-truth inflected form.", "label": "contrasting"} {"id": "test_272", "sentence1": "On the one hand, applying more than one iteration of additional lemma retrieval impacts the results only slightly, as those lemmas are assigned very small weights.", "sentence2": "we see performance differences > 2% between PCS-III-C and PCS-III-H for DEU, MLT, and SWE.", "label": "contrasting"} {"id": "test_273", "sentence1": "All these approaches first find a set of semantically similar sentences.", "sentence2": "finding isolated similar sentences are not enough to construct a dialog utterances' paraphrase.", "label": "contrasting"} {"id": "test_274", "sentence1": "Under the guidance of the reasoning chain, we learn a neural QG model to make the result satisfy the logical correspondence with the answer.", "sentence2": "the neural model is data-hungry, and the scale of training data mostly limits its performance.", "label": "contrasting"} {"id": "test_275", "sentence1": "Recently, the character-word lattice structure has been proved to be effective for Chinese named entity recognition (NER) by incorporating the word information.", "sentence2": "since the lattice structure is complex and dynamic, most existing lattice-based models are hard to fully utilize the parallel computation of GPUs and usually have a low inference-speed.", "label": "contrasting"} {"id": "test_276", "sentence1": "For example, queries \"red nike running shoes\", \"running nike shoes, red\" and \"red running shoes nike\" all refer to the same general product, despite differing in structure.", "sentence2": "item titles are structured, with brand, size, color, etc. all mentioned in a long sequence, which is also not how a conventional sentence is structured.", "label": "contrasting"} {"id": "test_277", "sentence1": "In this case, Q gen is similar to Q in that the item is somewhat related to Q gen , and there's a chance that I may be matched to Q gen due to keyword stuffing by sellers, or poor semantic matching.", "sentence2": "another mismatched query Q gen = pizza cutter is not a good candidate to generate, since it's highly unlikely that a reasonable search engine will show shoes for a query about pizza cutters.", "label": "contrasting"} {"id": "test_278", "sentence1": "Wu el al. (2019) applied dynamic convolutions using shared softmax-normalized filters of depth-wise on GLU-regulated inputs within a fixed reception field rather than global contexts, challenging the common self-attention-dominated intuition.", "sentence2": "all of the models, as mentioned earlier, adopt stacked CNNs rather than self-attention networks (SAN) to attend to the global contexts.", "label": "contrasting"} {"id": "test_279", "sentence1": "It can be also quantified by other manners, e.g. estimating the data likelihood with Monte Carlo approximation (Der Kiureghian and Ditlevsen, 2009) or validating the translation dis\u0002tribution using a well-trained NMT model (Zhang et al., 2018).", "sentence2": "to these time-consuming techniques, LM marginally increases the computational cost and easy to be applied, conforming to the original motivation of CL.", "label": "contrasting"} {"id": "test_280", "sentence1": "Since the speaker information is indispensable for coreference resolution, previous methods (Wiseman et al., 2016;Lee et al., 2017;Joshi et al., 2019a) usually convert the speaker information into binary features indicating whether two mentions are from the same speaker.", "sentence2": "we use a straightforward strategy that directly concatenates the speaker's name with the corresponding utterance.", "label": "contrasting"} {"id": "test_281", "sentence1": "Comparing with existing models (Lee et al., 2017Joshi et al., 2019b), the proposed question answering formalization has the flexibility of retrieving mentions left out at the mention proposal stage.", "sentence2": "since we still have the mention proposal model, we need to know in which situation missed mentions could be retrieved and in which situation they cannot.", "label": "contrasting"} {"id": "test_282", "sentence1": "The one-one target-source alignment 2(a) is the ideal condition of the projection.", "sentence2": "there could be many-to-one cases for the given words, leading to semantic role conflicts at the target language words.", "label": "contrasting"} {"id": "test_283", "sentence1": "Metrics which measure the word-level overlap like BLEU (Papineni et al., 2002) have been widely used for dialogue evaluation.", "sentence2": "these metrics do not fit into our setting well as we would like to diversify the response generation with an external corpus, the generations will inevitably differ greatly from the ground-truth references in the original conversational corpus.", "label": "contrasting"} {"id": "test_284", "sentence1": "The second class seeks to bring in extra information into existing corpus like structured knowledge (Zhao et al., 2018;Ghazvininejad et al., 2018;Dinan et al., 2019), personal information (Li et al., 2016b;Zhang et al., 2018a) or emotions (Shen et al., 2017b;Zhou et al., 2018).", "sentence2": "corpus with such annotations can be extremely costly to obtain and is usually limited to a specific domain with small data size.", "label": "contrasting"} {"id": "test_285", "sentence1": "The user requests to book one ticket in the second example, yet both HDSA and Human Response ask about the number once again.", "sentence2": "our model directly answers the questions with correct information.", "label": "contrasting"} {"id": "test_286", "sentence1": "One limitation of ReGAT (Li et al., 2019) lies in the fact that it solely consider the relations between objects in an image while neglect the importance of text information.", "sentence2": "our DC-GCN simultaneously capture visual relations in an image and textual relations in a question.", "label": "contrasting"} {"id": "test_287", "sentence1": "Figure 6a shows that in terms of validation perplexity, MDR and FB perform very similarly across target rates.", "sentence2": "figure 6b shows that at the end of training the difference between the target rate and the validation rate is smaller for MDR.", "label": "contrasting"} {"id": "test_288", "sentence1": "Previous conversational QA datasets provide the relevant document or passage that contain the answer of a query.", "sentence2": "in many real world scenarios such as FAQs, the answers need to be searched over the whole document collection.", "label": "contrasting"} {"id": "test_289", "sentence1": "For instance, all the span-based QA datasets, except CQ (Bao et al., 2016), contain more than 100k samples.", "sentence2": "the data size of most existing MCQA datasets are far less than 100k (see Table 1), and the smallest one only contains 660 samples.", "label": "contrasting"} {"id": "test_290", "sentence1": "On the one hand, technical proposals as pre-trained embeddings, finetuning, and end-to-end modeling, have advanced NLP greatly.", "sentence2": "neural advances often overlook MRL complexities, and disregard strategies that were proven useful for MRLs in the past.", "label": "contrasting"} {"id": "test_291", "sentence1": "Document-level information extraction requires a global understanding of the full document to annotate entities, their relations, and their saliency.", "sentence2": "annotating a scientific article is timeconsuming and requires expert annotators.", "label": "contrasting"} {"id": "test_292", "sentence1": "One common characteristic of most of the tasks is that the texts are not restricted to some rigid formats when generating.", "sentence2": "we may confront some special text paradigms such as Lyrics (assume the music score is given), Sonnet, SongCi (classical Chinese poetry of the Song dynasty), etc.", "label": "contrasting"} {"id": "test_293", "sentence1": "Our model can still generate high quality results on the aspects of format, rhyme as well as integrity.", "sentence2": "for corpus Sonnet, even though the model can generate 14 lines text, the quality is not as good as SongCi due to the insufficient training-set (only 100 samples).", "label": "contrasting"} {"id": "test_294", "sentence1": "Our relevance framework is partially inspired by the local components matching which we apply here to model the relevance of the components of the model's inputs.", "sentence2": "our work differs in several significant ways.", "label": "contrasting"} {"id": "test_295", "sentence1": "VisualBERT and CMR have a similar cross-modality alignment approach.", "sentence2": "visualBERT only uses the Transformer representations while CMR uses the relevance representations.", "label": "contrasting"} {"id": "test_296", "sentence1": "As we increase the number of layers in the visual Transformer and the cross-modality Transformer, it tends to improve accuracy.", "sentence2": "the performance becomes stable when there are more than five layers.", "label": "contrasting"} {"id": "test_297", "sentence1": "An agent needs to perform a functional communication task in a natural language (in this work, English).", "sentence2": "examples of linguistic communication about this functional task are not available -the only natural language data that can be used consist of examples of generic natural language, which are not grounded in the functional task.", "label": "contrasting"} {"id": "test_298", "sentence1": "Single-headed cross attention speeds up decoding: Despite removing learned self-attention from both the encoder and decoder, we did not observe huge efficiency or speed gains.", "sentence2": "reducing the source attention to just a single head results in more significant improvements.", "label": "contrasting"} {"id": "test_299", "sentence1": "We do find that the largest improvement in WinoMT accuracy consistently corresponds to the model predicting male and female entities in the closest ratio (see Appendix A).", "sentence2": "the best ratios for models adapted to these datasets are 2:1 or higher, and the accuracy improvement is small.", "label": "contrasting"} {"id": "test_300", "sentence1": "Ribeiro et al. (2018) test for comprehension of minimally modified sentences in an adversarial setup while trying to keep the overall semantics the same.", "sentence2": "we investigate large changes of meaning (negation) and context (mispriming).", "label": "contrasting"} {"id": "test_301", "sentence1": "Clustering of such short text streams has thus gained increasing attention in recent years due to many real-world applications like event tracking, hot topic detection, and news recommendation (Hadifar et al., 2019).", "sentence2": "due to the unique properties of short text streams such as infinite length, evolving patterns and sparse data representation, short text stream clustering is still a big challenge (Aggarwal et al., 2003;Mahdiraji, 2009).", "label": "contrasting"} {"id": "test_302", "sentence1": "The similarity-based text clustering approaches usually follow vector space model (VSM) to represent the cluster feature space (Din and Shao, 2020).", "sentence2": "a topic needs to be represented as the subspace of global feature space.", "label": "contrasting"} {"id": "test_303", "sentence1": "Human judges show surprisingly inferior performance on user profiling tasks, grounding their judgement in topical stereotypes (Carpenter et al., 2017).", "sentence2": "albeit more accurate thanks to capturing stylistic variation elements, statistical models are prone to stereotype propagation as well (Costa-jussa et al., 2019; Koolen and van Cranenburgh, 2017).", "label": "contrasting"} {"id": "test_304", "sentence1": "For the foreseeable future, legal decision-making will be the province of lawyers, not AI.", "sentence2": "one plausible use for MRC in a legal setting is as a screening tool for helping non-lawyers determine whether a case has enough merit to bother bringing in a lawyer.", "label": "contrasting"} {"id": "test_305", "sentence1": "Recent work has shown gains by improving the distribution of masked tokens , the order in which masked tokens are predicted (Yang et al., 2019), and the available context for replacing masked tokens (Dong et al., 2019).", "sentence2": "these methods typically focus on particular types of end tasks (e.g. span prediction, generation, etc.), limiting their applicability.", "label": "contrasting"} {"id": "test_306", "sentence1": "We aim, as much as possible, to control for differences unrelated to the pre-training objective.", "sentence2": "we do make minor changes to the learning rate and usage of layer normalisation in order to improve performance (tuning these separately for each objective).", "label": "contrasting"} {"id": "test_307", "sentence1": "Bidirectional encoders are crucial for SQuAD As noted in previous work (Devlin et al., 2019), just left-to-right decoder performs poorly on SQuAD, because future context is crucial in classification decisions.", "sentence2": "bART achieves similar performance with only half the number of bidirectional layers.", "label": "contrasting"} {"id": "test_308", "sentence1": "Unsurprisingly, model output is fluent and grammatical English.", "sentence2": "outputs are also highly abstractive, with few copied phrases.", "label": "contrasting"} {"id": "test_309", "sentence1": "One of the issues in the original ON-LSTM is that the master gates and the model-based importance score for each word are only conditioned on the word itself and the left context encoded in the previous hidden state.", "sentence2": "in order to infer the importance for a word in the overall sentence effectively, it is crucial to have a view over the entire sentence (i.e., including the context words on the right).", "label": "contrasting"} {"id": "test_310", "sentence1": "On the one hand, as GCN is directly dependent on the syntactic structures of the input sentences, it would not be able to learn effective representations for the sentences with new structures in the GCN-failure examples for RE.", "sentence2": "as CEON-LSTM only exploits a relaxed general form of the tree structures (i.e., the importance scores of the words), it will be able to generalize better to the new structures in the GCN-failure examples where the general tree form is still helpful to induce effective representations for RE.", "label": "contrasting"} {"id": "test_311", "sentence1": "ELMo down-samples the outputs of its convolutional layers by max-pooling over the feature maps.", "sentence2": "this operation is not ideal to adapt to new morphological patterns from other languages as the model tends to discard patterns from languages other than English.", "label": "contrasting"} {"id": "test_312", "sentence1": "The soft gazetteer features we propose instead take advantage of existing limited gazetteers and English knowledge bases using lowresource EL methods.", "sentence2": "to typical binary gazetteer features, the soft gazetteer feature values are continuous, lying between 0 and 1.", "label": "contrasting"} {"id": "test_313", "sentence1": "In addition, CGExpan-NoCN outperforms most baseline models, meaning that the pre-trained LM itself is powerful to capture entity similarities.", "sentence2": "it still cannot beat CGExpan-NoFilter model, which shows that we can properly guide the set expansion process by incorporating generated class names.", "label": "contrasting"} {"id": "test_314", "sentence1": "HINT proposes a ranking loss between humanbased importance scores (Das et al., 2016) and the gradient-based sensitivities.", "sentence2": "sCR does not require exact saliency ranks.", "label": "contrasting"} {"id": "test_315", "sentence1": "As observed by Selvaraju et al. (2019) and as shown in Fig. 2, we observe small improvements on VQAv2 when the models are fine-tuned on the entire train set.", "sentence2": "if we were to compare against the improvements in VQA-CPv2 in a fair manner, i.e., only use the instances with visual cues while fine-tuning, then, the performance on VQAv2 drops continuously during the course of the training.", "label": "contrasting"} {"id": "test_316", "sentence1": "In the original paper, Das et al. (2017) find that models which structurally encode dialog history, such as Memory Networks (Bordes et al., 2016) or Hierarchical Recurrent Encoders (Serban et al., 2017) improve performance.", "sentence2": "\"naive\" history modelling (in this case an encoder with late fusion/concatenation of current question, image and history encodings) might actually hurt performance.", "label": "contrasting"} {"id": "test_317", "sentence1": "In this paper, we show that competitive results on VisDial can indeed be achieved by replicating the top performing model for VQA (Yu et al., 2019b) -and effectively treating visual dialog as multiple rounds of question-answering, without taking history into account.", "sentence2": "we also show that these results can be significantly improved by encoding dialog history, as well as by fine-tuning on a more meaningful retrieval metric.", "label": "contrasting"} {"id": "test_318", "sentence1": "Other visual dialog tasks, such as GuessWhich? (Chattopadhyay et al., 2017) and GuessWhat?! (De Vries et al., 2017) take place in a goal-oriented setting, which according to Schlangen (2019), will lead to data containing more natural dialog phenomena.", "sentence2": "there is very limited evidence that dialog history indeed matters for these tasks (Yang et al., 2019).", "label": "contrasting"} {"id": "test_319", "sentence1": "The combined features approach models the relationships between the different features explicitly, but the large target spaces for morphologically rich languages further increase sparsity.", "sentence2": "separate feature modeling guarantees smaller target spaces for the individual features, but the hard separation between the features prevents modeling any interfeature dependencies.", "label": "contrasting"} {"id": "test_320", "sentence1": "The results are lower, both for MSA and EGY.", "sentence2": "the result for MSA is very close to the (Zalmout and Habash, 2017) baseline, which uses separate feature models (with the analyzer).", "label": "contrasting"} {"id": "test_321", "sentence1": "Although increases exist across all domains, these are most prominent in domains like TC (+5.36) that have a low density of named entities and where indomain models have access to limited amounts of data.", "sentence2": "the in-domain performance is better than the pooled method of training, which shows consistent drops in performance on some domains (-8.69 on WB, -6.77 on BC, -1.98 on CoNLL), where information from other domains did not benefit the model.", "label": "contrasting"} {"id": "test_322", "sentence1": "Our proposed model architecture takes 0.15 ms (33% increase) longer for inference than InDomain or PoolDomain models, which is a result of more model parameters.", "sentence2": "our proposed architecture is still 0.19 ms faster than using the InDomain+DomainClassifier approach.", "label": "contrasting"} {"id": "test_323", "sentence1": "State-of-the-art approaches for attribute value extraction (Zheng et al., 2018; Xu et al., 2019; Rezk et al., 2019) have employed deep learning to capture features of product attributes effectively for the extraction purpose.", "sentence2": "they are all designed without considering the product categories and thus cannot effectively capture the diversity of categories across the product taxonomy.", "label": "contrasting"} {"id": "test_324", "sentence1": "The CT and CAT models learn to map source code and natural language tokens into a joint embedding space such that semantically similar code-natural language pairs are projected to vectors that are close to each other.", "sentence2": "these two representations interact only in the final step when the global similarity of the sequence embeddings is calculated, but not during the first step when each sequence is encoded into its corresponding embedding.", "label": "contrasting"} {"id": "test_325", "sentence1": "Currently, the TClda are trained on student essays, while the TCpr only works on the source article.", "sentence2": "TCattn uses both student essays and the source article for TC generation.", "label": "contrasting"} {"id": "test_326", "sentence1": "Entity linking systems consider three sources of information: 1) similarity between mention strings and names for the KB entity; 2) comparison of the document con\u0002text to information about the KB entity (e.g. entity description); 3) information contained in the KB, such as entity popularity or inter-entity relations", "sentence2": "to the dense KBs in entity linking, concept linking uses sparse ontologies, which contain a unique identifier (CUI), title, and links to synonyms and related concepts, but rarely longform text.", "label": "contrasting"} {"id": "test_327", "sentence1": "We ran experiments that padded the names with synonyms or other forms of available text within the knowledge base.", "sentence2": "we did not see consistent improvements.", "label": "contrasting"} {"id": "test_328", "sentence1": "Depending on the application, a less accurate but faster linker might be a better choice (e.g. for all clinical notes at a medical institution).", "sentence2": "a more complex linker, such as ours, maybe a better option for specific subsets of notes that require better accuracy (e.g., the results of specific clinical studies).", "label": "contrasting"} {"id": "test_329", "sentence1": "In natural language processing, Natural Language Inference (NLI)-a task whereby a system determines whether a pair of sentences instantiates in an entailment, a contradiction, or a neutral relation-has been useful for training and evaluating models on sentential reasoning.", "sentence2": "linguists and philosophers now recognize that there are separate semantic and pragmatic modes of reasoning (Grice, 1975; Clark, 1996; Beaver, 1997; Horn and Ward, 2004; Potts, 2015), and it is not clear which of these modes, if either, NLI models learn", "label": "contrasting"} {"id": "test_330", "sentence1": "For example, the Independent model will produce identical scores for each output label, if it chooses to completely ignore the input explanations.", "sentence2": "the model is still free to learn a different kind of bias which is an outcome of the fact that natural language explanations convey ideas through both content and form.", "label": "contrasting"} {"id": "test_331", "sentence1": "The results demonstrate a much weaker link between NILE-NS's predictions and associated explanations.", "sentence2": "nILE behaves more expectedly.", "label": "contrasting"} {"id": "test_332", "sentence1": "This led folk wisdom to suggest that modeling higher-order features in a neural parser would not bring additional advantages, and nearly all recent research on dependency parsing was restricted to first-order models (Dozat and Manning, 2016; Smith et al., 2018a). Kulmizev et al. (2019) further reinforced this belief comparing transition and graph-based decoders (but none of which higher order); Falenska and Kuhn (2019) suggested that higher-order features become redundant because the parsing models encode them implicitly.", "sentence2": "there is some evidence that neural parsers still benefit from structure modeling.", "label": "contrasting"} {"id": "test_333", "sentence1": "This approach, named adversarial training (AT), has been reported to be highly effective on image classification (Goodfellow et al., 2015), text classification (Miyato et al., 2017), as well as sequence labeling (Yasunaga et al., 2018).", "sentence2": "aT is limited to a supervised scenario, which uses the labels to compute adversarial losses.", "label": "contrasting"} {"id": "test_334", "sentence1": "To apply the conventional VAT on a model with CRF, one can calculate the KL divergence on the label distribution of each token between the original examples and adversarial examples.", "sentence2": "it is sub-optimal because the transition probabilities are not taken into account.", "label": "contrasting"} {"id": "test_335", "sentence1": "VAT achieved state-of-the-art performance for image classification tasks (Miyato et al., 2019), and proved to be more efficient than traditional semi-supervised approaches, such as entropy minimization (Grandvalet and Bengio, 2004) and selftraining (Yarowsky, 1995), from a recent study (Oliver et al., 2018).", "sentence2": "despite the successful applications on text classification (Miyato et al., 2017), VAT has not shown great benefits to semi-supervised sequence labeling tasks, due to its incompatibility with CRF.", "label": "contrasting"} {"id": "test_336", "sentence1": "With multiple layers, SpellGCN can aggregate the information in more hops and therefore, achieve better performance.", "sentence2": "the F1score drops when the number of layers is larger than 3.", "label": "contrasting"} {"id": "test_337", "sentence1": "The evident way to construct a corpus with NL questions and their corresponding OT queries would consist of two main parts: first, collect a set of NL questions, and then create the corresponding OT queries to these questions.", "sentence2": "this approach is very time-consuming and has a major issue.", "label": "contrasting"} {"id": "test_338", "sentence1": "On the other hand, LC-QuaD 2.0 contains an average of 2 hops (equivalent to two joins in relational databases) per query, which lies in the nature of graph database queries that are optimized for handling queries that range over multiple triple patterns.", "sentence2": "lC-QuaD 2.0 lacks complexity when considering more complex components (e.g., Group By, Set-Operation, etc.).", "label": "contrasting"} {"id": "test_339", "sentence1": "Table 2 is a comparison of existing English MWP corpora.", "sentence2": "these existing corpora are either limited in terms of the diversity of the associated problem types (as well as lexicon usage patterns), or lacking information such as difficulty levels.", "label": "contrasting"} {"id": "test_340", "sentence1": "Likewise, we could also enlarge the training-set by duplicating MWPs without affecting the CLD value against the test-set.", "sentence2": "it would be also meaningless as no new information would be provided.", "label": "contrasting"} {"id": "test_341", "sentence1": "Because this type of evaluation is typically task-specific, it can be conducted in multilingual settings.", "sentence2": "training a range of task-specific multilingual models might require significant resources, namely, training time and computational power.", "label": "contrasting"} {"id": "test_342", "sentence1": "Typelevel probing tasks have the advantage of containing less bias (domain, annotator, and majority class); whereas token-level tests might be sensitive to the domain biases from the underlying full-text data.", "sentence2": "token-level tests have the advantage of being more lexically diverse; whereas type-level tasks can be less diverse for some languages like Spanish, French, and English.", "label": "contrasting"} {"id": "test_343", "sentence1": "Unlike previously introduced embedding models, ELMo provides contextualized embeddings, that is, the same words would have different representations when used in different contexts.", "sentence2": "our probing tests are type-level (as opposed to token-level), thus we only use the representations generated independently per each token both for the intrinsic and extrinsic experiments.", "label": "contrasting"} {"id": "test_344", "sentence1": "Another potential reason for the difference in ranking is the domain of the data underlying the respective data sets: For the majority of the languages, POS, DEP, and SRL data originates from the same treebanks and has gold (expert) annotations.", "sentence2": "nER and XnLI data sets are generally compiled from a different, and often diverse set of resources.", "label": "contrasting"} {"id": "test_345", "sentence1": "The abstract syntax can be vastly different in domains ranging from mathematics to tourist phrasebooks.", "sentence2": "the linguistic mechanisms needed in the concrete syntax-morphology, agreement, word order-are largely the same in all areas of discourse.", "label": "contrasting"} {"id": "test_346", "sentence1": "For this purpose, it is enough to express all desired content in one way: One does not need to cover all possible ways to express things.", "sentence2": "in wide-coverage parsing, this is a serious limitation.", "label": "contrasting"} {"id": "test_347", "sentence1": "This led to a series of extensions as described in Section 4.5, first meant to cover the missing English structures.", "sentence2": "if the ultimate goal is to build an interlingual grammar, the structures designed for English are not necessarily adequate for other languages-in particular, they might not allow for compositional linearizations.", "label": "contrasting"} {"id": "test_348", "sentence1": "There, every synset corresponds to one abstract function and then the function's linearization in each language produces all words in the language as variants.", "sentence2": "a more detailed analysis shows that this is not ideal.", "label": "contrasting"} {"id": "test_349", "sentence1": "Our focus in this paper has been on recognizing valid chains of reasoning, assuming a retrieval step that retrieves a reasonable pool of candidates to start with (Section 3.2).", "sentence2": "the retrieval step itself is not perfect: For QASC, designed so that at least one valid chain always exists, the retrieved pool of 10 contains no valid chains for 24% of the questions (upper bound in Table 2), capping the overall system's performance.", "label": "contrasting"} {"id": "test_350", "sentence1": "Multitask learning (Caruana, 1997;Collobert and Weston, 2008) seeks to learn a single model that can solve multiple tasks simultaneously, similar to our framework that seeks to learn a model that can solve many tasks.", "sentence2": "in multitask learning each task is learned from examples, and the model is not able to generalize to unseen tasks.", "label": "contrasting"} {"id": "test_351", "sentence1": "Attribution of natural disasters/collective misfortune is a widely-studied political science problem.", "sentence2": "such studies typically rely on surveys, expert opinions, or external signals such as voting outcomes.", "label": "contrasting"} {"id": "test_352", "sentence1": "For instance, the most-recent PEW survey (Pew) focused on India was conducted in 2018 on only 2,521 users.", "sentence2": "our data set consists of comments from 43,859 users.", "label": "contrasting"} {"id": "test_353", "sentence1": "We observe that, on the detection task, all the BERT based models perform similarly.", "sentence2": "on the resolution task, the F1 score substantially improves as we keep adding sophistication to our model architecture.", "label": "contrasting"} {"id": "test_354", "sentence1": "Third, a more comprehensive evaluation methodology would consider both the exact-match accuracy and the execution-match accuracy, because two logic forms can be semantically equivalent yet do not match precisely in their surface forms.", "sentence2": "as shown in Table 1, most existing work is only evaluated with the exact-match accuracy.", "label": "contrasting"} {"id": "test_355", "sentence1": "Consequently, the execution engines of domain-specific MRs need to be significantly customized for different domains, requiring plenty of manual efforts.", "sentence2": "sQL is a domain-general MR for querying relational databases.", "label": "contrasting"} {"id": "test_356", "sentence1": "There is a predicate tomorrow in all three domainspecific MRs, and this predicate can directly align to the description in the utterance.", "sentence2": "one needs to explicitly express the concrete date values in the SQL query; this requirement can be a heavy burden for neural approaches, especially when the values will change over time.", "label": "contrasting"} {"id": "test_357", "sentence1": "Unfortunately, this integral is intractable due to the complex relationship between X and Z.", "sentence2": "related latent variable models like variational autoencoders (VAEs; Kingma and Welling (2013)) learn by optimizing a variational lower bound on the log marginal likelihood.", "label": "contrasting"} {"id": "test_358", "sentence1": "For instance, we could feed both sentences into the semantic encoder and pool their representations.", "sentence2": "in practice we find that alternating works well and also can be used to obtain sentence embeddings for text that is not part of a translation pair.", "label": "contrasting"} {"id": "test_359", "sentence1": "This model is similar to Infersent (Conneau et al., 2017) in that it is trained on natural language inference data, SNLI (Bowman et al., 2015).", "sentence2": "instead of using pretrained word embeddings, they fine-tune BERT in a way to induce sentence embeddings.", "label": "contrasting"} {"id": "test_360", "sentence1": "Since BGT W/O LANGVARS also has significantly better performance on these tasks, most of this gain seems to be due to the prior having a regularizing effect.", "sentence2": "bGT outperforms bGT W/O LANGVARS overall, and we hypothesize that the gap in performance between these two models is due to bGT being able to strip away the language-specific information in the representations with its languagespecific variables, allowing for the semantics of the sentences to be more directly compared.", "label": "contrasting"} {"id": "test_361", "sentence1": "Japanese is a very distant language to English both in its writing system and in its sentence structure (it is an SOV language, where English is an SVO language).", "sentence2": "despite these difference, the semantic encoder strongly outperforms the English language-specific encoder, suggesting that the underlying meaning of the sentence is much better captured by the semantic encoder.", "label": "contrasting"} {"id": "test_362", "sentence1": "We can observe that the inherent data imbalance problem also exists in MAVEN.", "sentence2": "as MAVEN is large-scale, 41% and 82% event types have more than 500 and 100 instances respectively.", "label": "contrasting"} {"id": "test_363", "sentence1": "Most recently, continuous improvements have been achieved by combining multiple kinds of information in KGs or using more sophisticated embedding models.", "sentence2": "the performances of most approaches are still not satisfactory.", "label": "contrasting"} {"id": "test_364", "sentence1": "To generate the PCG of two KGs, we can first pair all the entities from two KGs as nodes, and then use Equation 1 to generate edges between nodes.", "sentence2": "kGs usually contain large number of entities, the PCG of two large-scale kGs will contain huge number of nodes.", "label": "contrasting"} {"id": "test_365", "sentence1": "Recent studies on single-document summarization (SDS) benefit from the advances in neural sequence learning (Nallapati et al., 2016;See et al., 2017;Chen and Bansal, 2018;Narayan et al., 2018) as well as pretrained language models (Liu and Lapata, 2019;Lewis et al., 2019;Zhang et al., 2020) and make great progress.", "sentence2": "in multi-document summarization (MDS) tasks, neural models are still facing challenges and often underperform classical statistical methods built upon handcrafted features (Kulesza and Taskar, 2012).", "label": "contrasting"} {"id": "test_366", "sentence1": "One extension (Cho et al., 2019) of these studies uses capsule networks (Hinton et al., 2018) to improve redundancy measures.", "sentence2": "its capsule networks are pre-trained on SDS and fixed as feature inputs of classical methods without end-to-end representation learning.", "label": "contrasting"} {"id": "test_367", "sentence1": "Compared to hard cutoff, our soft attention favors top-ranked candidates of the sentence ranker (MMR).", "sentence2": "it does not discard low-ranked ones, as the ranker is imperfect, and those sentences ranked low may also contribute to a high-quality summary.", "label": "contrasting"} {"id": "test_368", "sentence1": "RL-MMR has a more salient and non-redundant summary, as it is end-to-end trained with advances in SDS for sentence representation learning while maintaining the benefits of classical MDS approaches.", "sentence2": "mmR alone only considers lexical similarity; The redundancy mea-sure in DPP-Caps-Comb is pre-trained on one SDS dataset with weak supervision and fixed during the training of DPP.", "label": "contrasting"} {"id": "test_369", "sentence1": "The authors considered several predefined probe architectures and picked one of them based on a manually defined criterion.", "sentence2": "the variational code gives probe architecture as a byproduct of training and does not need human guidance.", "label": "contrasting"} {"id": "test_370", "sentence1": "To this day, scientific publications still serve as a fundamental fixed-domain benchmark for neural KPE methods (Meng et al., 2017;Alzaidy et al., 2019;Sahrawat et al., 2019) due to the availability of ample data of this kind.", "sentence2": "experiments have revealed that KPE methods trained directly on such corpora do not generalize well to other web-related genres or other types of documents (Chen et al., 2018;Xiong et al., 2019), where there may be far more heterogeneity in topics, content and structure, and there may be more variation in terms of where a key phrase may appear.", "label": "contrasting"} {"id": "test_371", "sentence1": "Case #2 shows a similar situation where the model with visual features finds the proper keyphrases that are much larger in font size, while the text-only model selects nouns elsewhere.", "sentence2": "case #3 demonstrates a typical kind of web page where visual features can be misleading: an indexing page.", "label": "contrasting"} {"id": "test_372", "sentence1": "As a result, the relationships between entities are not captured.", "sentence2": "since KB is naturally a graph structure (nodes are entities and edges are relations between entities).", "label": "contrasting"} {"id": "test_373", "sentence1": "Moreover, structural knowledge such as dependency relationships has recently been investigated on some tasks (e.g., relation extraction) (Peng et al., 2017;Song et al., 2018) and shown to be effective in the model's generalizability.", "sentence2": "such dependency relationships (essentially also graph structure) have not been explored in dialogue systems, again missing great potential for improvements.", "label": "contrasting"} {"id": "test_374", "sentence1": "different predecessors in H ) should have different impacts on the output hidden state h t , and we expect our model to capture that.", "sentence2": "the inputs may have different number of predecessors at different timesteps.", "label": "contrasting"} {"id": "test_375", "sentence1": "We can observe that our model without the graph encoder has a 1.6% absolute value loss (over 25% in ratio) in BLEU score and a 1.1% absolute value loss (9.8% in ratio) in entity F1 on MultiWOZ 2.1, which suggests that the overall quality of the generated sentences are better improved by our graph encoder.", "sentence2": "ours without knowledge graph means that we do not use the graph structure to store and retrieve the external knowledge data.", "label": "contrasting"} {"id": "test_376", "sentence1": "These systems rely on offline (batch) training and have drawn recent criticism due to their inability to adapt to new contexts (Linzen, 2020).", "sentence2": "humans acquire language from evolving environments, require a small memory footprint (Mc-Clelland et al., 1995), and can generalize their knowledge to newer tasks (Sprouse et al., 2013).", "label": "contrasting"} {"id": "test_377", "sentence1": "However, these works are mainly designed for image classification tasks where the training data has \"clear\" task boundaries-i.e., training stream are partitioned into disjoint subsequences.", "sentence2": "task boundaries in VisCOLL are unknown and \"smooth\" (i.e., with gradual transitions between tasks)-a setting that is closer to real-world situations.", "label": "contrasting"} {"id": "test_378", "sentence1": "In particular, Li et al. (2020) study a closely related task of continual learning of sequence prediction for synthetic instruction following", "sentence2": "their techniques for separating semantics and syntax is restricted to text-only case.", "label": "contrasting"} {"id": "test_379", "sentence1": "Prior work uses only the article content and metadata including title, date, domain, and authors.", "sentence2": "news articles often contain photos and captions as well.", "label": "contrasting"} {"id": "test_380", "sentence1": "Despite the fact that leveraging metadata significantly improves the performance of Grover, it also appears that the accuracy does not vary much with the exclusion of different types of metadata.", "sentence2": "we observe a surprising observation that leveraging all metadata causes the detection accuracy to decrease.", "label": "contrasting"} {"id": "test_381", "sentence1": "Todd Frazier's sacrifice fly accounted for the first run before Jose Bautista drove in the next two with a line drive RBI single to right, and a bases-loaded single by Todd Frazier also scored a run.", "sentence2": "deJong and Bader homered off Bobby Wahl to begin the Cardinals' comeback.", "label": "contrasting"} {"id": "test_382", "sentence1": "Each sub-problem is worthy of being standardized and continually studied given a well defined objective and data sets so that the performance could be fairly evaluated and the progress can be continually made.", "sentence2": "it is not easy in the current methodology, since each pipeline's strategies are closely bonded to own implementation.", "label": "contrasting"} {"id": "test_383", "sentence1": "Finetuning a pretrained language model (Dai and Le, 2015;Howard and Ruder, 2018) often delivers competitive performance partly because pretraining leads to a better initialization across various downstream tasks than training from scratch (Hao et al., 2019).", "sentence2": "finetuning on individual NLP tasks is not parameter-efficient.", "label": "contrasting"} {"id": "test_384", "sentence1": "Mallya et al. (2018) explicitly update weights in a task-specific classifier layer.", "sentence2": "we show that end-to-end learning of selective masks, consistently for both the pretrained language model and a randomly initialized classifier layer, achieves good performance.", "label": "contrasting"} {"id": "test_385", "sentence1": "Large-scale training datasets lie at the core of the recent success of neural machine translation (NMT) models.", "sentence2": "the complex patterns and potential noises in the large-scale data make training NMT models difficult.", "label": "contrasting"} {"id": "test_386", "sentence1": "Neural machine translation (NMT) is a data-hungry approach, which requires a large amount of data to train a well-performing NMT model (Koehn and Knowles, 2017).", "sentence2": "the complex patterns and potential noises in the large-scale data make training NMT models difficult.", "label": "contrasting"} {"id": "test_387", "sentence1": "Another stream is to schedule the order of training examples according to their difficulty, e.g., curriculum learning which has been applied to the training of NMT models successfully (Kocmi and Bojar, 2017; Zhang et al., 2018; Platanios et al., 2019; Liu et al., 2020b).", "sentence2": "we explore strategies to simplify the difficult (i.e., inactive) examples without changing the model architecture and model training strategy.", "label": "contrasting"} {"id": "test_388", "sentence1": "For the latter, NMT models tend to prefer a more typical alternative to a relatively rare but correct one (e.g., French \"Il\" is often wrongly translated to the more common \"it\" than \"he\" ).", "sentence2": "However, these seemingly trivial errors can erode translation to the extent that they can be easily distinguishable from human-translated texts (Laubli et al. \u00a8 , 2018).", "label": "contrasting"} {"id": "test_389", "sentence1": "Both of the discriminative losses essentially promote the probability of the positive (i.e., correct) sample.", "sentence2": "the intuition behind using the additional loss over the standard loss is that the fine-tuning here focuses on improving the positive sample over the negative sample that the model has learnt to produce, rather than over the entire probability distribution over the full vocabulary.", "label": "contrasting"} {"id": "test_390", "sentence1": "In the second example, we observe a biased anticipation case where the NMT system had to emit a wrong translation chien ('dog') before seeing the noun 'bird'.", "sentence2": "the multimodal model successfully leveraged the visual context for anticipation and correctly handled the adjective-noun placement phenomenon.", "label": "contrasting"} {"id": "test_391", "sentence1": "Our approach is most similar to Bjerva et al. (2019a), as they build a generative model from typological features and use language embeddings, extracted from factored language modelling at character-level, as a prior of the model to extend the language coverage.", "sentence2": "our method primarily differs as it is mainly based in linear algebra, encodes information from both sources since the beginning, and can deal with a small number of shared entries (e.g. 23 from LW) to compute robust representations.", "label": "contrasting"} {"id": "test_392", "sentence1": "We work with 53 languages pre-processed by (Qi et al., 2018), from where we mapped the ISO 639-1 codes to the ISO 693-2 standard.", "sentence2": "we need to manually correct the mapping of some codes to identify the correct language vector in the URIEL (Littell et al., 2017) library: \u2022 zh (zho , Chinese macro-language) mapped to cmn (Mandarin Chinese).", "label": "contrasting"} {"id": "test_393", "sentence1": "In German, we reach the maximum 0.41 when the number of words in each topic equals 2, and the minimum when it equals 100.", "sentence2": "we observe the most noticeable changes when we vary the number of topics in French (Ousidhoum et al., 2019) such that B 1 = 0.34 when |T| = 2 versus 0.21 when |T| = 7 and back to 0.37 when |T| = 100.", "label": "contrasting"} {"id": "test_394", "sentence1": "On the other hand, we observe the most noticeable changes when we vary the number of topics in French (Ousidhoum et al., 2019) such that B 1 = 0.34 when |T| = 2 versus 0.21 when |T| = 7 and back to 0.37 when |T| = 100.", "sentence2": "we remark overall cohesion despite the change in topic numbers especially in the case of Italian and Portuguese caused by the limited numbers of search keywords, that equal 5 and 7 respectively.", "label": "contrasting"} {"id": "test_395", "sentence1": "Waseem and Hovy (2016), Founta et al. (2018) and Ousidhoum et al. (2019) report using different keywords and hashtags to collect tweets.", "sentence2": "the scores shown in Table 4 indicate that the datasets might carry similar meanings, specifically because WUP relies on hypernymy rather than common vocabulary use.", "label": "contrasting"} {"id": "test_396", "sentence1": "On the other hand, the copy model (row 3) significantly improves the BLEU scores by 36.2-37.6 points, by learning to re-use words in input texts 4 .", "sentence2": "it still suffers the small data size, and its outputs are worse than the original questions without any transformation (row 1).", "label": "contrasting"} {"id": "test_397", "sentence1": "For system-level evaluation, metrics which can use the reference translations for quality estimation, such as BLEU, generally achieved consistently high correlation with human evaluation for all language pairs.", "sentence2": "qE models (including our qE model and submitted systems for the qE as a Metric task) are not allowed to use the reference translations for quality estimation and tend to generate more unstable results: high correlation with human evaluation for some language pairs but very low or even negative Pearson correlation with human evaluation for some other language pairs.", "label": "contrasting"} {"id": "test_398", "sentence1": "As a sequence-to-sequence generation task, neural machine translation (NMT) naturally contains intrinsic uncertainty, where a single sentence in one language has multiple valid counterparts in the other.", "sentence2": "the dominant methods for NMT only observe one of them from the parallel corpora for the model training but have to deal with adequate variations under the same meaning at inference.", "label": "contrasting"} {"id": "test_399", "sentence1": "Recently, there are increasing number of studies investigating the effects of quantifying uncertainties in different applications Kendall and Gal, 2017;Xiao and Wang, 2018;Zhang et al., 2019b,a;Shen et al., 2019).", "sentence2": "most work in NMT has focused on improving accuracy without much consideration for the intrinsic uncertainty of the translation task itself.", "label": "contrasting"} {"id": "test_400", "sentence1": "when h tends to 0 our controlled sampling method achieves lowest BLEU scores but highest edit distances.", "sentence2": "if we increase h gradually, it can be quickly simplified to greedy search.", "label": "contrasting"} {"id": "test_401", "sentence1": "A sequence-to-sequence (seq2seq) learning with neural networks empirically shows to be an effective framework for grammatical error correction (GEC), which takes a sentence with errors as input and outputs the corrected one.", "sentence2": "the performance of GEC models with the seq2seq framework heavily relies on the size and quality of the corpus on hand.", "label": "contrasting"} {"id": "test_402", "sentence1": "The former applies text editing operations such as substitution, deletion, insertion and shuffle, to introduce noises into original sentences, and the latter trains a clean-to-noise model for error generation.", "sentence2": "the noise-corrupted sentences are often poorly readable, which are quite different from those made by humans.", "label": "contrasting"} {"id": "test_403", "sentence1": "Once a vulnerable position is determined, the token at that position is usually replaced with one of its synonyms.", "sentence2": "generating adversarial examples through such synonym-based replacement is no longer applicable to the GEC task.", "label": "contrasting"} {"id": "test_404", "sentence1": "Adversarial training by means of adding the adversarial examples into the training set can effectively improve the models' robustness.", "sentence2": "some studies show that the models tend to overfit the noises, and the accuracy of the clean data will drop if the number of adversarial examples dominates the training set.", "label": "contrasting"} {"id": "test_405", "sentence1": "Indeed, LDA implicitly assumes that \u03a8\u00a8 = Unif(1, .., K) deterministically-i.e., that every topic is assumed a priori to contain the same number of tokens.", "sentence2": "the HDP model learns this distribution from the data by letting \u03a8 \u223c GEM(\u03b3).", "label": "contrasting"} {"id": "test_406", "sentence1": "Semisupervised methods that utilize such external corpora have been successful in English STS.", "sentence2": "the need for external corpora is a major obstacle when applying STS, a fundamental technology, to low-resource languages.", "label": "contrasting"} {"id": "test_407", "sentence1": "One particularly promising usage of BERT-based models for unsupervised STS is BERTScore , which was originally proposed as an automatic evaluation metric.", "sentence2": "our preliminary experiments 17 show that BERTScore performs poorly on unsupervised STS.", "label": "contrasting"} {"id": "test_408", "sentence1": "From the figure, we find that overall the average accuracy raises when K increases from 2 to 8, which suggests the importance of disentangling components.", "sentence2": "when K grows larger than 8, the performance starts to decline.", "label": "contrasting"} {"id": "test_409", "sentence1": "As shown in Figure 6(b), except for the case of n = 1, the other settings have comparable performance.", "sentence2": "it can be seen that when n = 4, the average accuracy on the last task is the highest, which indicates that the model has the strongest ability to avoid catastrophic forgetting problem when n = 4.", "label": "contrasting"} {"id": "test_410", "sentence1": "The subproblems are separately solved using existing techniques.", "sentence2": "existing unsupervised multilingual approaches (Chen and Cardie, 2018;Heyman et al., 2019;Alaux et al., 2019) solve the above subproblems jointly.", "label": "contrasting"} {"id": "test_411", "sentence1": "We also observe that UMWE fails at mapping Dutch language embeddings in the multilingual space even though Dutch is close to English.", "sentence2": "in a separate bilingual experiment, UMWE learns an effective English-Dutch crosslingual space (obtaining an average en-nl and nl-en score of 75.2).", "label": "contrasting"} {"id": "test_412", "sentence1": "We observe that the proposed SL-GeoMM learns a highly effective multilingual space and obtains the best overall result, illustrating its robustness in this challenging setting.", "sentence2": "other multilingual approaches fail to learn a reasonably good multilingual space.", "label": "contrasting"} {"id": "test_413", "sentence1": "In recent years, pre-trained language models, such as GPT (Radford et al., 2018), BERT (Devlin et al., 2018), XL-Net (Yang et al., 2019), have been proposed and applied to many NLP tasks, yielding state-of-the-art performances.", "sentence2": "the promising results of the pre-trained language models come with the high costs of computation and memory in inference, which obstruct these pre-trained language models to be deployed on resource-constrained devices and real-time applications.", "label": "contrasting"} {"id": "test_414", "sentence1": "Concretely, both the embedding-layer distillation and the prediction-layer distillation employ the one-to-one layer mapping as in TinyBERT and BERT-PKD, where the two student layers are guided by the corresponding teacher layers, respectively.", "sentence2": "different from the previous works, we propose to exploit the many-to-many layer mapping for Transformer (intermediate lay-ers) distillation (attention-based distillation and hidden states based distillation), where each student attention layer (resp. hidden layers).", "label": "contrasting"} {"id": "test_415", "sentence1": "In this paper, we compare our BERT-EMD with several state-of-the-art BERT compression approaches, including the original 4/6-layer BERT models (Devlin et al., 2018), DistilBERT (Tang et al., 2019), BERT-PKD , Tiny-BERT (Jiao et al., 2019), BERT-of-Theseus (Xu et al., 2020).", "sentence2": "the original TinyBERT employs a data augmentation strategy in the training process, which is different from the other baseline models.", "label": "contrasting"} {"id": "test_416", "sentence1": "Previous works have shown that as the number of retrieved passages increases, so does the performance of the reader.", "sentence2": "they assume all retrieved passages are of equal importance and allocate the same amount of computation to them, leading to a substantial increase in computational cost.", "label": "contrasting"} {"id": "test_417", "sentence1": "Open-Domain Question Answering (ODQA) requires a system to answer questions using a large collection of documents as the information source.", "sentence2": "to context-based machine comprehension, where models are to extract answers from single paragraphs or documents, it poses a fundamental technical challenge in machine reading at scale (Chen et al., 2017) .", "label": "contrasting"} {"id": "test_418", "sentence1": "Models for reading comprehension (RC) commonly restrict their output space to the set of all single contiguous spans from the input, in order to alleviate the learning problem and avoid the need for a model that generates text explicitly.", "sentence2": "forcing an answer to be a single span can be restrictive, and some recent datasets also include multi-span questions, i.e., questions whose answer is a set of non-contiguous spans in the text.", "label": "contrasting"} {"id": "test_419", "sentence1": "When the answer spans appear only once in the input, this is simple, since the ground-truth tagging is immediately available.", "sentence2": "there are many cases where a given answer span appears multiple times in the input.", "label": "contrasting"} {"id": "test_420", "sentence1": "For query generation, prior research has focused mostly on extending standard Seq2Seq models where the input is a concatenation of earlier queries a user has submitted in a session (Sordoni et al.,Figure 1: An example search session where a user issues queries and optionally performs clicking at timestamps 1 to n. At time n+1, the user issues q n+1 following the previous search context of length n. 2015; Dehghani et al., 2017).", "sentence2": "literature often leaves out the influence of clickthrough actions (i.e., red blocks in Figure 1), which we argue should be taken into account in the generative process as they could be surrogates of the user's implicit search intent (Yin et al., 2016).", "label": "contrasting"} {"id": "test_421", "sentence1": "One of the widely exercised steps to establish a semantic understanding of social media is Entity Linking (EL), i.e., the task of linking entities within a text to a suitable concept in a reference Knowledge Graph (KG) (Liu et al., 2013;Yang and Chang, 2015;Yang et al., 2016; Ran et al., 2018).", "sentence2": "it is well-documented that poorly composed contexts, the ubiquitous presence of colloquialisms, shortened forms, typing/spelling mistakes, and out-of-vocabulary words introduce challenges for effective utilisation of social media text (Baldwin et al., 2013;Michel and Neubig, 2018).", "label": "contrasting"} {"id": "test_422", "sentence1": "As noted by Ethayarajh (2019) the deeper BERT goes, the more \"contextualized\" its representation becomes.", "sentence2": "interpreting semantics of entities requires contextual knowledge in different degrees and always taking the last layer's output may not be the best solution.", "label": "contrasting"} {"id": "test_423", "sentence1": "Compared to similar corpora, COMETA has the largest scale.", "sentence2": "from a learning perspective the lack of sufficient regularity in the data could still leave its toll at test phase.", "label": "contrasting"} {"id": "test_424", "sentence1": "Much of the recent progress in NLP is due to the transfer learning paradigm in which Transformerbased models first try to learn task-independent linguistic knowledge from large corpora, and then get fine-tuned on small datasets for specific tasks.", "sentence2": "these models are overparametrized: we now know that most Transformer heads and even layers can be pruned without significant loss in performance (Voita et al., 2019;Kovaleva et al., 2019;Michel et al., 2019).", "label": "contrasting"} {"id": "test_425", "sentence1": "There are recent works (Nguyen et al., 2019; Agarwal et al., 2020) also applying the Transformer to model the interactions among many entities.", "sentence2": "their models neglect the important early interaction of the answer entity and cannot naturally leverage the pretrained language representations from BERT like ours.", "label": "contrasting"} {"id": "test_426", "sentence1": "In the example, both the baseline model and the model with the sub-instruction module completes the task successfully.", "sentence2": "unlike the baseline model which fails to follow the instruction and stops within 3 meters of the target by chance, our model correctly identifies the completeness of each sub-instruction, guides the agent to walk on the described path and eventually stops right at the target position.", "label": "contrasting"} {"id": "test_427", "sentence1": "Although replacing a real user with a user simulator could address the issue, the simulator only roughly approximates real user statistics, and its development process is costly (Su et al., 2016).", "sentence2": "humans could independently reason potential responses based on past experiences from the true environment.", "label": "contrasting"} {"id": "test_428", "sentence1": "We propose a model-agnostic approach, COPT, that can be applied to any adversarial learning-based dialogue generation models.", "sentence2": "to existing approaches, it learns on counterfactual responses inferred from the structural causal model, taking advantage of observed responses.", "label": "contrasting"} {"id": "test_429", "sentence1": "As an important research issue in the natural language processing community, multi-label emotion detection has been drawing more and more attention in the last few years.", "sentence2": "almost all existing studies focus on one modality (e.g., textual modality).", "label": "contrasting"} {"id": "test_430", "sentence1": "This implies that the success of previous models may over-rely on the confounding non-target aspects, but not necessarily on the target aspect only.", "sentence2": "no datasets can be used to analyze the aspect robustness more in depth.", "label": "contrasting"} {"id": "test_431", "sentence1": "These two ratios should ideally be both 400%, because there are three generation strategies, plus one original sentence.", "sentence2": "this gap is because not every original test sentence can qualify for every generation strategy.", "label": "contrasting"} {"id": "test_432", "sentence1": "That is to say, whether one sentence could be selected depends on its salience and the redundancy with other selected sentences.", "sentence2": "it is still difficult to model the dependency exactly.", "label": "contrasting"} {"id": "test_433", "sentence1": "Neural models have achieved remarkable success on relation extraction (RE) benchmarks.", "sentence2": "there is no clear understanding which type of information affects existing RE models to make decisions and how to further improve the performance of these models.", "label": "contrasting"} {"id": "test_434", "sentence1": "From the observations in Section 2, we know that both context and entity type information is beneficial for RE models.", "sentence2": "in some cases RE models cannot well understand the relational patterns in context and rely on the shallow cues of entity mentions for classification.", "label": "contrasting"} {"id": "test_435", "sentence1": "Alt et al. (2020) also point out that there may exist shallow cues in entity mentions.", "sentence2": "there have not been systematical analyses about the topic and to the best of our knowledge, we are the first one to thoroughly carry out these studies.", "label": "contrasting"} {"id": "test_436", "sentence1": "Other models selected operands first before constructing expression trees with operators in the second step (Roy et al., 2015; Roy and Roth, 2015).", "sentence2": "such two-step procedures in these early attempts can be performed via a single-step procedure with neural models.", "label": "contrasting"} {"id": "test_437", "sentence1": "It has received significant attention in question answering systems for structured data (Wang et al., 2015; Zhong et al., 2017; Yu et al., 2018b; Xu et al., 2020).", "sentence2": "training a semantic parser with good accuracy requires a large amount of annotated data, which is expensive to acquire.", "label": "contrasting"} {"id": "test_438", "sentence1": "Neural networks have the merits of convenient end-to-end training and good generalization, however, they typically need a lot of training data and are not interpretable.", "sentence2": "logicbased expert systems are interpretable and require less or no training data.", "label": "contrasting"} {"id": "test_439", "sentence1": "A unique feature of this operationalisation of lexical ambiguity is that it is language independent.", "sentence2": "the quality of a possible approximation will vary from language to language, depending on the models and the data available in that language.", "label": "contrasting"} {"id": "test_440", "sentence1": "If p(m | w) is concentrated in a small region of the meaning space (corresponding to a word with nuanced implementations of the same sense), the bound in eq. (13) could be relatively tight.", "sentence2": "a word with several unrelated homophones would correspond to a highly structured p(m | w) (e.g. with multiple modes in far distant regions of the space) for which this normal approximation would result in a very loose upper bound.", "label": "contrasting"} {"id": "test_441", "sentence1": "Deep learning has led to significant improvement in text summarization with various methods investigated and improved ROUGE scores reported over the years.", "sentence2": "gaps still exist between summaries produced by automatic summarizers and human professionals.", "label": "contrasting"} {"id": "test_442", "sentence1": "In the research literature, human evaluation has been conducted as a complement (Narayan et al., 2018).", "sentence2": "human evaluation reports that accompany ROUGE scores are limited in scope and coverage.", "label": "contrasting"} {"id": "test_443", "sentence1": "The above methods assign one score to each summarization output.", "sentence2": "to these methods, our errorcount based metrics are motivated by MQM for human writing, and are more fine-grained and informative.", "label": "contrasting"} {"id": "test_444", "sentence1": "On PolyTope, as a representative of abstractive models, BART overwhelmingly outperforms the others (p < 0.01 using t-test).", "sentence2": "excluding BART, extractive models take the following top three places.", "label": "contrasting"} {"id": "test_445", "sentence1": "With respect to Accuracy, extractive methods are notably stronger in terms of Inacc Intrinsic and Extrinsic, which reflects that through directly copying snippets from the source, extractive methods are guaranteed to produce a summary with fair grammaticality, rationality and loyalty.", "sentence2": "extractive methods do not show stronger performances in Addition and Omission, which is because extracted sentences contain information not directly relevant to the main points.", "label": "contrasting"} {"id": "test_446", "sentence1": "There is a high proportion in the first five sentences and a smooth tail over all positions for reference summaries.", "sentence2": "bertSumExt and SummaRuN-Ner extract sentences mostly from the beginning, thereby missing useful information towards the end.", "label": "contrasting"} {"id": "test_447", "sentence1": "Compared with Point-Generator, Point-Generator-with-Coverage reduces Duplication errors from 68 to 11 and Omission errors from 286 to 256, proving that coverage is useful for better content selection.", "sentence2": "point-Generator-with-Coverage yields more Addition and Inacc Intrinsic errors than point-Generator.", "label": "contrasting"} {"id": "test_448", "sentence1": "As can be seen, abstractive models tend to neglect sentences in the middle and at the end of source documents (e.g., Bottom-Up, BertSumExtAbs), indicating that performance of abstractive summarizers is strongly affected by the leading bias of dataset.", "sentence2": "bART can attend to sentences all around the whole document, slightly closer to the distribution of golden reference.", "label": "contrasting"} {"id": "test_449", "sentence1": "Their main goal is to verify the faithfulness and factuality in abstractive models.", "sentence2": "we evaluate both rule-based baselines and extractive/abstractive summarizers on 8 error metrics, among which faithfulness and factuality are included.", "label": "contrasting"} {"id": "test_450", "sentence1": "It has been conjectured that multilingual information can help monolingual word sense disambiguation (WSD).", "sentence2": "existing WSD systems rarely consider multilingual information, and no effective method has been proposed for improving WSD by generating translations.", "label": "contrasting"} {"id": "test_451", "sentence1": "Our first method extends the idea of Apidianaki and Gong (2015) to constrain S(e) based on sensetranslation mappings in BabelNet.", "sentence2": "instead of relying on a single translation, we incorporate multiple languages by taking the intersection of the individual sets of senses; that is, we rule out senses if their corresponding BabelNet synsets do not contain translations from all target languages.", "label": "contrasting"} {"id": "test_452", "sentence1": "Supervised systems are trained on sense-annotated corpora and generally outperform knowledge-based systems.", "sentence2": "knowledge-based systems usually apply graph-based algorithms to a semantic network and thus do not require any sense-annotated corpora.", "label": "contrasting"} {"id": "test_453", "sentence1": "The ability to fuse sentences is highly attractive for summarization systems because it is an essential step to produce succinct abstracts.", "sentence2": "to date, summarizers can fail on fusing sentences.", "label": "contrasting"} {"id": "test_454", "sentence1": "Fusing two sentences together coherently requires connective phrases and sometimes requires rephrasing parts of sentences.", "sentence2": "higher abstraction does not mean higher quality fusions, especially in neural models.", "label": "contrasting"} {"id": "test_455", "sentence1": "Recent innovations in Transformer-based ranking models have advanced the state-ofthe-art in information retrieval.", "sentence2": "these Transformers are computationally expensive, and their opaque hidden states make it hard to understand the ranking process.", "label": "contrasting"} {"id": "test_456", "sentence1": "One could argue that this is a superficial problem, as we can always give the model more free bits and decrease the loss in intermediary positions.", "sentence2": "this is not so simple because increasing capacity leads to a worse model fit, as was noted by Alemi et al. (2018).", "label": "contrasting"} {"id": "test_457", "sentence1": "While Yang et al. (2017) and Kim et al. (2018) both consider the use of pretrained LMs as encoders, the weights are not frozen such that it is hard to disentangle the impact of pretraining from subsequent training.", "sentence2": "we freeze the weights so that the effect of pretraining can not be overridden.", "label": "contrasting"} {"id": "test_458", "sentence1": "Both baselines and variants have roughly similarly high agreement.", "sentence2": "our variants produce more diverse beginnings, while still managing to reproduce the topic or sentiment of the original document.", "label": "contrasting"} {"id": "test_459", "sentence1": "Finally, verb-shuffling focuses on verbs as the salient element of an event, and should teach both principles of verb ordering and of verb suitability for context, and avoid artifacts from reordering arguments.", "sentence2": "since verbs are shuffled naively, the task can in some cases be too easy due to differences in verb selectional preferences.", "label": "contrasting"} {"id": "test_460", "sentence1": "In the vision and machine learning community, unsupervised induction of structured image representations (aka scene graphs or world models) has been receiving increasing attention (Eslami et al., 2016; Burgess et al., 2019; Kipf et al., 2020).", "sentence2": "they typically rely solely on visual signal.", "label": "contrasting"} {"id": "test_461", "sentence1": "Previous work has shown that incorporating natural language explanation into the classification training loop is effective in various settings (Andreas et al., 2018; Mu et al., 2020).", "sentence2": "previous work neglects the fact that there is usually a limited time budget to interact with domain experts (e.g., medical experts, biologists) and high-quality natural language explanations are expensive, by nature.", "label": "contrasting"} {"id": "test_462", "sentence1": "For example, an image with a Ringbilled gull has the description: \"This is a white bird with a grey wing and orange eyes and beak.\"", "sentence2": "this description also fits perfectly with a California gull (Figure 1).", "label": "contrasting"} {"id": "test_463", "sentence1": "Our results indicate that too many parameters can also harm multilinguality.", "sentence2": "in practice it is difficult to create a model with so many parameters that it is overparameterized when being trained on 104 Wikipedias.", "label": "contrasting"} {"id": "test_464", "sentence1": "One might argue that our model 17 in Table 1 of the main paper is simply not trained enough and thus not multilingual.", "sentence2": "table 10 shows that even when continuing to train this model for a long time no multilinguality arises.", "label": "contrasting"} {"id": "test_465", "sentence1": "This suggests that multilingual models can stimulate positive transfer for low-resource languages when monolingual models overfit.", "sentence2": "when we compare bilingual models on English, models trained using different sizes of fr/ru data obtain similar performance, indicating that the training size of the source language has little impact on negative interference on the target language (English in this case).", "label": "contrasting"} {"id": "test_466", "sentence1": "Unlike language-specific adapters that can hinder transferability, shared adapters improve both within-language and cross-lingual performance with the extra capacity.", "sentence2": "meta adapters still obtain better performance.", "label": "contrasting"} {"id": "test_467", "sentence1": "They observe that having more languages results in better zero-shot performance.", "sentence2": "several artifacts arise, as described by Dabre et al. (2020); Zhang et al. (2020); Aharoni et al. (2019); Arivazhagan et al. (2019), like off\u0002target translation and insufficient modeling capac\u0002ity of the MNMT models.", "label": "contrasting"} {"id": "test_468", "sentence1": "This might indicate that -at least during the adaptation -important information is captured in the encoder's adapter layer (in line with previous reports by Kudugunta et al., 2019) or that the decoder adaptation grows dependent on the encoder adapters, to the point where dropping the latter degrades the system.", "sentence2": "further analysis would be needed to confirm either of these hypotheses.", "label": "contrasting"} {"id": "test_469", "sentence1": "Not only is beam search usually more accurate than greedy search, but it also outputs a diverse set of decodings, enabling reranking approaches to further improve accuracy (Yee et al., 2019; Ng et al., 2019; Charniak and Johnson, 2005; Ge and Mooney, 2006).", "sentence2": "it is challenging to optimize the performance of beam search for modern neural architectures.", "label": "contrasting"} {"id": "test_470", "sentence1": "FIXED-OURS is slower than Fairseq's implementation.", "sentence2": "while the two implementations achieve more similar BLEU on the development set, FIXED-OURS achieves higher BLEU on the test set (49.75 vs 49.57 on De-En and 39.19 vs 38.98 on Ru-En).", "label": "contrasting"} {"id": "test_471", "sentence1": "The produced meaning representations can then potentially be used to improve downstream NLP applications (e.g., Issa et al., 2018;Song et al., 2019;Mihaylov and Frank, 2019), though the introduction of large pretrained language models has shown that explicit formal meaning representations might not be a necessary component to achieve high accuracy.", "sentence2": "it is now known that these models lack reasoning capabilities, often simply exploiting statistical artifacts in the data sets, instead of actually understanding language (Niven and Kao, 2019;McCoy et al., 2019).", "label": "contrasting"} {"id": "test_472", "sentence1": "A possible advantage of this model is that it might handle longer sentences and documents better.", "sentence2": "it might be harder to tune (Popel and Bojar, 2018) 2 and its improved performance has mainly been shown for large data sets, as opposed to the generally smaller semantic parsing data sets (Section 3.3).", "label": "contrasting"} {"id": "test_473", "sentence1": "For both methods, it results in a clear and significant improvement over the BERT-only baseline, 87.6 versus 88.1.", "sentence2": "another common method of improving performance is adding linguistic features to the tokenlevel representations.", "label": "contrasting"} {"id": "test_474", "sentence1": "This was done for efficiency and memory purposes, it did not make a difference in terms of F1-score.", "sentence2": "for the Transformer model this improved F1-score by around 0.5.", "label": "contrasting"} {"id": "test_475", "sentence1": "Data efficiency can be improved by optimizing pre-training directly for future fine-tuning with few examples; this can be treated as a meta-learning problem.", "sentence2": "standard meta-learning techniques require many training tasks in order to generalize; unfortunately, finding a diverse set of such supervised tasks is usually difficult.", "label": "contrasting"} {"id": "test_476", "sentence1": "Bansal et al. (2019) proposed an approach that applies to diverse tasks to enable practical meta-learning models and evaluate on generalization to new tasks.", "sentence2": "they rely on supervised task data from multiple tasks and suffer from meta-overfitting as we show in our empirical results.", "label": "contrasting"} {"id": "test_477", "sentence1": "Owing to the warp layers, our training time per step and the GPU memory footprint is lower than LEOPARD (Bansal et al., 2019).", "sentence2": "our training typically runs much longer as the model doesn't overfit unlike LEOPARD (see learning rate trajectory in main paper).", "label": "contrasting"} {"id": "test_478", "sentence1": "As Wiki itself is a collaborative knowledge repository, editors are likely to attack others due to disputes on specific domain knowledge.", "sentence2": "the users are the general public who post comments and tweets more casually for Yahoo and Twitter.", "label": "contrasting"} {"id": "test_479", "sentence1": "Multilingual contextual embeddings have demonstrated state-of-the-art performance in zero-shot cross-lingual transfer learning, where multilingual BERT is fine-tuned on one source language and evaluated on a different target language.", "sentence2": "published results for mBERT zero-shot accuracy vary as much as 17 points on the MLDoc classification task across four papers.", "label": "contrasting"} {"id": "test_480", "sentence1": "Many models inspired by BERT have since surpassed its performance.", "sentence2": "in contrast to the original BERT paper, many obtained better results by excluding the NSP task.", "label": "contrasting"} {"id": "test_481", "sentence1": "The decoupled biLSTM extended with ELMo inputs is able to outperform the transformer model initialised with RoBERTa pretraining.", "sentence2": "the best performance is achieved by using the transformer model with BART-large pretraining, with the decoupled model fine-tuned jointly on top of it (Lewis et al., 2019).", "label": "contrasting"} {"id": "test_482", "sentence1": "The concept of Dialogue Act (DA) is universal across different task-oriented dialogue domains-the act of \"request\" carries the same speaker intention whether it is for restaurant reservation or flight booking.", "sentence2": "dA taggers trained on one domain do not generalize well to other domains, which leaves us with the expensive need for a large amount of annotated data in the target domain.", "label": "contrasting"} {"id": "test_483", "sentence1": "It is often challenging and costly to obtain a large amount of in-domain dialogues with annotations.", "sentence2": "unlabeled dialogue corpora in target domain can easily be curated from past conversation logs or collected via crowd-sourcing (Byrne et al., 2019;Budzianowski et al., 2018) at a more reasonable cost.", "label": "contrasting"} {"id": "test_484", "sentence1": "In prior work (Xie et al., 2019;Wei and Zou, 2019), unsupervised data augmentation methods including word replacement and backtranslation have been shown useful for short written text classification.", "sentence2": "such augmentation methods are shown to be less effective (Shleifer, 2019) when used with pre-trained models.", "label": "contrasting"} {"id": "test_485", "sentence1": "We find that for both tense and mood in the Indo-Aryan family, our model identifies required-agreement primarily for conjoined verbs, which mostly need to agree only if they share the same subject.", "sentence2": "subsequent analysis revealed that in the treebanks nearly 50% of the agreeing verbs do not share the same subject but do agree by chance.", "label": "contrasting"} {"id": "test_486", "sentence1": "Professor Tanja Kallio and doctoral candidate Sami Tuomi consider the realisation of this goal entirely possible.", "sentence2": "\"scientifically we are in the dark about the consequences of rewilding, and we worry about the general lack of critical thinking surrounding these often very expensive attempts at conservation.", "label": "contrasting"} {"id": "test_487", "sentence1": "Such pairs show consistent improvement (+5 to +10), which suggests that the model learns to align the parallel knowledge from the source language to the target language.", "sentence2": "we also must note that the effect is strongly dependent on the size of the overlapping sets.", "label": "contrasting"} {"id": "test_488", "sentence1": "Students need good textbooks to study before they can pass an exam, and the same holds for a good machine reading model.", "sentence2": "finding the information needed to answer a question, especially for questions in such a narrow domain as the subjects studied in high schools, usually requires a collection of specialized texts.", "label": "contrasting"} {"id": "test_489", "sentence1": "However, little work has looked further up the pipeline and relied on the assumption that biases in data originate in human cognition.", "sentence2": "this assumption motivates our work: an unsupervised approach to detecting implicit gender bias in text.", "label": "contrasting"} {"id": "test_490", "sentence1": "Psychology studies often examine human perceptions through word associations (Greenwald et al., 1998).", "sentence2": "the implicit nature of bias suggests that human annotations for bias detection may not be reliable, which motivates an unsupervised approach.", "label": "contrasting"} {"id": "test_491", "sentence1": "is likely addressed towards a woman and identify it as biased.", "sentence2": "we only want the model to learn that references to appearance are indicative of gender if they occur in unsolicited contexts.", "label": "contrasting"} {"id": "test_492", "sentence1": "For example, humans need the supervision of what is a noun before they do POS tagging, or what is a tiger in Wordnet before they classify an image of tiger in ImageNet.", "sentence2": "for NLI, people are able to entail that a A man plays a piano contradicts b A man plays the clarinet for his family without any supervision from the NLI labels.", "label": "contrasting"} {"id": "test_493", "sentence1": "Both depGCN and kumaGCN can correctly classify the sentiment of \"service\" as negative.", "sentence2": "depGCN cannot recognize the positive sentiment of \"atmosphere\" while kumaGCN can.", "label": "contrasting"} {"id": "test_494", "sentence1": "For the target \"atmosphere\", depGCN assigns the highest weight to the word \"terrible\", which is an irrelevant sentiment word to this target, leading to an incorrect prediction.", "sentence2": "our model assigns the largest weight to the key sentiment word \"cozy\", classifying it correctly.", "label": "contrasting"} {"id": "test_495", "sentence1": "Although collecting tags from users is time\u0002consuming and often suffers from coverage issues (Katakis et al., 2008), NLP techniques like those in Kar et al. (2018b) and Gorinski and Lapata (2018) can be employed to generate tags automatically from written narratives such as synopses.", "sentence2": "existing supervised approaches suffer from two significant weaknesses.", "label": "contrasting"} {"id": "test_496", "sentence1": "August is selected to perform the rhapsody he's been composing at the same concert.", "sentence2": "wizard, who found out about August's performance by Arthur, interrupts the rehearsal and claims to be his father, and manages to pull August out of the school.", "label": "contrasting"} {"id": "test_497", "sentence1": "Mozart would be an absolute imbecile compared to this little kid August Rush, and for those familiar with music, this aspect (the foundation, really) just kills the movie.It is impossible to play like Michael Hedges in your first few minutes with a guitar.", "sentence2": "i just finished watching August Rush and i am in no way exaggerating when i say that it is by far the best movie i have ever seen.", "label": "contrasting"} {"id": "test_498", "sentence1": "Im not sure who would actually enjoy this movie, maybe if you're 70, or under 12 but for everyone else I'd save your time.The acting itself wasn't bad, though the more interesting characters were played by Terrence Howard and Robin Williams, and they were both severely under-developed as you wanted to know more about them and less about this kid with the stupid smile all the time..", "sentence2": "while the movie has a modern setting, it shares many plot elements with OLIVER TWIST, ending even better.It begins with a young couple of musicians that meets and has a one-night stand, and when she becomes pregnant her dad does everything to make her believe that the child died at birth, although he just put the child for adoption.A decade later the boy, Evan, lives in a orphanage and is mocked by the other kids because of his talents in music, that makes him like a savant with powerful skills.", "label": "contrasting"} {"id": "test_499", "sentence1": "Although this film plays well to a broad audience, it is very mystical and based on simple, yet emotional themes that will play flat to some movie-goers.If you have strong parental feelings or enjoy movies centered on the power of human love and attraction, this story will move you like few films ever have.", "sentence2": "if you are easily bored with themes that are lacking in danger and suspense or prefer gritty true-to-life movies, this one may come off as a disappointment.The screenplay seems written as a spiritual message intimating that there is an energy field that connects all of life, and music is one of the domains available to any who care to experience it.The plot is simple but deep in implication-an orphaned boy wants to reunite with his parents and feels that his inherited musical genius can somehow guarantee their return.", "label": "contrasting"} {"id": "test_500", "sentence1": "Graph convolutional networks (GCN) is demonstrated to be an effective approach to model such contextual information between words in many NLP tasks (Marcheggiani and Titov, 2017;Huang and Carley, 2019;De Cao et al., 2019); thus we want to determine whether this approach can also help CCG supertagging.", "sentence2": "we cannot directly apply conventional GCN models to CCG supertagging because in most of the previous studies the GCN models are built over the edges in the dependency tree of an input sentence.", "label": "contrasting"} {"id": "test_501", "sentence1": "Having more training data can help reduce overfitting and improve model robustness.", "sentence2": "preparing a large amount of annotated data is usually costly, labor intensive and time-consuming.", "label": "contrasting"} {"id": "test_502", "sentence1": "In this work, we will focus on denoising recurrent neural network autoencoders (Vincent et al., 2010;Shen et al., 2020; see Appendix A).", "sentence2": "any advancement in this research direction will directly benefit our framework.", "label": "contrasting"} {"id": "test_503", "sentence1": "This observation may seem counter to the widely seen success of finetuning across other NLP scenarios, in particular with pretrained transformer models like BERT (Devlin et al., 2019).", "sentence2": "finetuning does not always lead to better performance.", "label": "contrasting"} {"id": "test_504", "sentence1": "Recent work starts to use gradient (Michel et al., 2019;Ebrahimi et al., 2017) to guide the search for universal trigger (Wallace et al., 2019) that are applicable to arbitrary sentences to fool the learner, though the reported attack success rate is rather low or they suffer from inefficiency when applied to other NLP tasks.", "sentence2": "our proposed T3 framework is able to effectively generate syntactically correct adversarial text, achieving high targeted attack success rates across different models on multiple tasks.", "label": "contrasting"} {"id": "test_505", "sentence1": "There are also systems (Luo et al., 2018; Kumar et al., 2019) that incorporate sense definitions into language models and achieve state-of-the-art performance.", "sentence2": "most of the systems are implemented in a supervised manner using a widely exploited sense-annotated corpus, SemCor (Miller et al., 1994), and merging knowledge from the sense inventory as a supplement.", "label": "contrasting"} {"id": "test_506", "sentence1": "By using regex-based extractors and a list of comprehensive dictionaries that capture crucial domain vocabularies, LUSTRE can generate rules that achieve SoTA results.", "sentence2": "for more complex and realistic scenarios, dictionaries may not be available and regex-based extractors alone are not expressive enough.", "label": "contrasting"} {"id": "test_507", "sentence1": "EMR and LwF can achieve competitive performance at the beginning.", "sentence2": "the gap between the two baselines and our method KCN becomes wider as more new classes arrive.", "label": "contrasting"} {"id": "test_508", "sentence1": "Very few models exist that can predict either open vocab (Rashkin et al., 2018), or variable size output .", "sentence2": "no existing task has both open vocabulary and variable-size low specificity-placing OPENPI in a novel space.", "label": "contrasting"} {"id": "test_509", "sentence1": "The SCoNE dataset (Long et al., 2016) contains paragraphs describing a changing world state in three synthetic, deterministic domains.", "sentence2": "approaches developed using synthetic data often fail to handle the inherent complexity in language when applied to organic, real-world data (Hermann et al., 2015;Winograd, 1972).", "label": "contrasting"} {"id": "test_510", "sentence1": "For informativeness, we notice that all models perform well on the seen domains.", "sentence2": "on unseen domains, the Naive approach fares poorly.", "label": "contrasting"} {"id": "test_511", "sentence1": "Very similar to our task, Kang et al. (2019) developed language models informed by discourse relations on the bridging task; given the first and last sen\u0002tences, predicting the intermediate sentences (bidirectional flow).", "sentence2": "they did not explicitly predict content words given context nor use them as a self-supervision signal in training.", "label": "contrasting"} {"id": "test_512", "sentence1": "As an alternative metric of attention explainablity, (Jain and Wallace, 2019) considers the relationship between attention weights and gradient-based feature importance score of each word.", "sentence2": "prior research suggests using word as a unit of importance feature is rather artificial, as word is contextualized by, and interacts with other words: (Wiegreffe and Pinter, 2019) observes such limitation, and Shapley (Chen et al., 2018) measures interaction between features for capturing dependency of arbitrary subsets.", "label": "contrasting"} {"id": "test_513", "sentence1": "The GNN-based models are particularly strong in this setting (see Appendix C), and this suggests that transferring knowledge about the relevancy of facts from structured to unstructured models may be a promising direction.", "sentence2": "at the same time, the improvements for generalization were less substantial, indicating that some reasoning capacities are difficult to distill in this manner.", "label": "contrasting"} {"id": "test_514", "sentence1": "This is not surprising as the generalization ability is a known issue in modern NLP models and is an ongoing research topic (Bahdanau et al., 2019; Andreas, 2019).", "sentence2": "the generalization is in parallel with our contribution that is to improve the reasoning ability of NLP models.", "label": "contrasting"} {"id": "test_515", "sentence1": "Predictive methods such as probing are flexible: Any task with data can be assessed.", "sentence2": "they only track predictability of pre-defined categories, limiting their descriptive power.", "label": "contrasting"} {"id": "test_516", "sentence1": "Since nPMI is information-theoretic and chance-corrected, it is a reliable indicator of the degree of information about gold labels contained in a set of predicted clusters.", "sentence2": "it is relatively insensitive to cluster granularity (e.g., the total number of predicted categories, or whether a single gold category is split into many different predicted clusters), which is better understood through our other metrics.", "label": "contrasting"} {"id": "test_517", "sentence1": "On one hand, this reflects surface patterns: primary core arguments are usually close to the verb, with ARG0 on the left and ARG1 on the right; trailing arguments and modifiers tend to be prepositional phrases or subordinate clauses; and modals and negation are identified by lexical and positional cues.", "sentence2": "this also reflects error patterns in state-of-the-art systems, where label errors can sometimes be traced to ontological choices in PropBank, which distinguish between arguments and adjuncts that have very similar meaning Kingsbury et al., 2002).", "label": "contrasting"} {"id": "test_518", "sentence1": "Most of the existing work has adopted static sentiment lexicons as linguistic resource (Qian et al., 2017; Chen et al., 2019), and equipped each word with a fixed sentiment polarity across different contexts.", "sentence2": "the same word may play different sentiment roles in different contexts due to the variety of part-ofspeech tags and word senses.", "label": "contrasting"} {"id": "test_519", "sentence1": "We observe that our proposed MT-H-LSTM-CRF consistently outperforms the baseline models.", "sentence2": "it performs slightly worse on RR-submission than on RR-passage, plausibly because there is no context information (i.e., background knowledge from original submissions) shared between different passage pairs.", "label": "contrasting"} {"id": "test_520", "sentence1": "Sentence [1] is non-hyperbolic because the fact that \"her one step equals my two steps\" is not anything that would be surprising to anyone.", "sentence2": "if one changes the number from \"two\" to \"100\", then the resulting sentence becomes hyperbolic because in reality it is not possible that one person's step would equal another person's 100 steps.", "label": "contrasting"} {"id": "test_521", "sentence1": "Their proposal is similar to ours; they exclude attention weights that do not affect the output owing to the application of transformation f and input x in the analysis.", "sentence2": "our proposal differs from theirs in some aspects.", "label": "contrasting"} {"id": "test_522", "sentence1": "The target word in (1 a) is associated 3 with the gold gloss (1 b) from WordNet (Fellbaum, 1998), the most used sense inventory in WSD.", "sentence2": "generationary arguably provides a better gloss (1 c).", "label": "contrasting"} {"id": "test_523", "sentence1": "It is not surprising that machine learning methods can easily surpass human performance if sufficient data is available (Wang et al., 2018).", "sentence2": "data acquisition is a challenging task for some special domains.", "label": "contrasting"} {"id": "test_524", "sentence1": "The core idea of our method is finding a different entity for intervening on an entity in the observational example.", "sentence2": "finding a new entity set in a specific domain needs human efforts to collect entities, which has no difference from annotating more data.", "label": "contrasting"} {"id": "test_525", "sentence1": "Second, the training and test data in these benchmarks are sampled from the same corpus, and therefore the training data usually have high mention coverage on the test data, i.e., a large proportion of mentions in the test set have been observed in the training set.", "sentence2": "it is obvious that this high coverage is inconsistent with the primary goal of NER models, which is expected to identify unseen mentions from new data by capturing the generalization knowledge about names and contexts.", "label": "contrasting"} {"id": "test_526", "sentence1": "This is because they only annotate named mentions but ignore nominal and pronominal mentions.", "sentence2": "the context of named and nominal/pronominal mentions is generally identical, and therefore the models will be unable to distinguish between them once name regularity is removed.", "label": "contrasting"} {"id": "test_527", "sentence1": "Temporal KGs often exhibit multiple simultaneous non-Euclidean structures, such as hierarchical and cyclic structures.", "sentence2": "existing embedding approaches for temporal KGs typically learn entity representations and their dynamic evolution in the Euclidean space, which might not capture such intrinsic structures very well.", "label": "contrasting"} {"id": "test_528", "sentence1": "More recently, generalized manifolds of constant curvature to a product manifold combining hyperbolic, spherical, and Euclidean components.", "sentence2": "these methods consider graph data as static models and lack the ability to capture temporally evolving dynamics.", "label": "contrasting"} {"id": "test_529", "sentence1": "Building upon recent NLI systems, our approach leverages representations from unsupervised pretraining, and finetunes a multiclass classifier over the BERT model (Devlin et al., 2019).", "sentence2": "we first consider other models for related tasks.", "label": "contrasting"} {"id": "test_530", "sentence1": "Such results are problematic for entailment (since it is defined to depend on the truth of the premise).", "sentence2": "our problem is primarily about the meaning of answers.", "label": "contrasting"} {"id": "test_531", "sentence1": "The approach achieves consistently better performance compared to the first two rows on the inter-domain structure prediction task (For both, original Parseval and RST-Parseval), as we have previously shown in Huber and Carenini (2019).", "sentence2": "only considering two out of three nuclearity classes (N-S and S-N), the system performs rather poorly on the nuclearity classification task.", "label": "contrasting"} {"id": "test_532", "sentence1": "Our proposed new method (BERT-BASE-LWAN) that employs LWAN on top of BERT-BASE has the best results among all methods on EURLEX57K and AMAZON13K, when all and frequent labels are considered.", "sentence2": "in both datasets, the results are comparable to BERT-BASE, indicating that the multi-head attention mechanism of BERT can effectively handle the large number of labels.", "label": "contrasting"} {"id": "test_533", "sentence1": "To identify whether an example is biased, they employ a shallow model f b , a simple model trained to directly compute p(y|b(x)), where the features b(x) are hand-crafted based on the task-specific knowledge of the biases.", "sentence2": "obtaining the prior information to design b(x) requires a dataset-specific analysis (Sharma et al., 2018).", "label": "contrasting"} {"id": "test_534", "sentence1": "Training a shallow model The analysis suggests that we can obtain a substitute f b by taking a checkpoint of the main model early in the training, i.e., when the model has only seen a small portion of the training data.", "sentence2": "we observe that the resulting model makes predictions with rather low confidence, i.e., assigns a low probability to the predicted label.", "label": "contrasting"} {"id": "test_535", "sentence1": "This indicates that unregularized training optimizes faster on certain examples, possibly due to the presence of biases.", "sentence2": "self-debiased training maintains relatively less variability of losses throughout the training.", "label": "contrasting"} {"id": "test_536", "sentence1": "Closely related to our work, Singh et al. (2019) showed that replacing segments of the training data with their translation during fine-tuning is helpful.", "sentence2": "they attribute this behavior to a data augmentation effect, which we believe should be reconsidered given the new evidence we provide.", "label": "contrasting"} {"id": "test_537", "sentence1": "For example, by framing the immigration issue using the morality frame or using the security frame, the reader is primed to accept the liberal or conservative perspectives, respectively.", "sentence2": "as shown in Example 1, in some cases this analysis is too coarse grained, as both articles frame the issue using the economic frame, suggesting that a finer grained analysis is needed to capture the differences in perspective.", "label": "contrasting"} {"id": "test_538", "sentence1": "In Example 1, both texts use the Economic frame using same unigram indicator ('wage').", "sentence2": "other words in the text can help identify the nuanced talking points (e.g., 'minimum wage' in case of left and 'stagnant wages' in case of right).", "label": "contrasting"} {"id": "test_539", "sentence1": "Availability of large-scale datasets has enabled the use of statistical machine learning in vision and language understanding, and has lead to significant advances.", "sentence2": "the commonly used evaluation criterion is the performance of models on test-samples drawn from the same distribution as the training dataset, which cannot be a measure of generalization.", "label": "contrasting"} {"id": "test_540", "sentence1": "The goal of OOD generalization is to mitigate negative bias while learning to perform the task.", "sentence2": "existing methods such as LMH (Clark et al., 2019) try to remove all biases between question-answer pairs, by penalizing examples that can be answered without looking at the image; we believe this to be counterproductive.", "label": "contrasting"} {"id": "test_541", "sentence1": "Notice that both mutations do not significantly change the input, most of the pixels in the image and words in the question are unchanged, and the type of reasoning required to answer the question is unchanged.", "sentence2": "the mutation significantly changes the answer.", "label": "contrasting"} {"id": "test_542", "sentence1": "As expected, the batch-aware strategies, DAL and Core-Set, which were designed to increase diversity, are characterized by the most diverse batches, with DAL achieving the highest diversity values, demonstrating the success of using mini-queries (Gissin and Shalev-Shwartz, 2019) to reduce redundancy of the selected examples.", "sentence2": "the other strategies tend to select less diverse batches, i.e., they are prone to choose redundant examples, especially in the imbalanced-practical scenario.", "label": "contrasting"} {"id": "test_543", "sentence1": "This roughly translates to increasing the throughput of the training process.", "sentence2": "when performing inference on a single data point, the latency of making predictions seems to dominate the runtime (Jouppi et al., 2017).", "label": "contrasting"} {"id": "test_544", "sentence1": "Recent work has shown that contextualised embeddings pre-trained on large written corpora can be fine-tuned on smaller spoken language corpora to learn structures of spoken language (Tran et al., 2019).", "sentence2": "for NLP tasks, fillers and all disfluencies are typically removed in pre-processing, as NLP models achieve highest accuracy on syntactically correct utterances.", "label": "contrasting"} {"id": "test_545", "sentence1": "An assumption one could make based on the work by Radford et al. (2019), is that with this model, the results for any further downstream task would be improved by the presence of fillers.", "sentence2": "we observe that to predict the persuasiveness of the speaker (using the high level attribute of persuasiveness annotated in the dataset (Park et al., 2014)), following the same procedure as outlined in subsubsection 2.1.2, that fillers, in fact, are not a discriminative feature.", "label": "contrasting"} {"id": "test_546", "sentence1": "Stehwien and Vu (2017) and Stehwien et al. (2018) (henceforth, SVS18) showed that neural methods can perform comparably to traditional methods us\u0002ing a relatively small amount of speech context\u2014 just a single word on either side of the target word.", "sentence2": "since pitch accents are deviations from a speaker\u2019s average pitch, intensity, and duration, we hypothesize that, as in some non-neural mod\u0002els (e.g. Levow 2005; Rosenberg and Hirschberg 2009), a wider input context will allow the model to better determine the speaker\u2019s baseline for these features an", "label": "contrasting"} {"id": "test_547", "sentence1": "Recent advances in deep learning present a promising prospect in multimodal stock forecasting by analyzing online news , and social media (Guo et al., 2018) to learn latent patterns affecting stock prices (Jiang, 2020).", "sentence2": "the challenging aspect in stock forecasting is that most existing work treats stock movements to be independent of each other, contrary to true market function (Diebold and Y\u0131lmaz, 2014).", "label": "contrasting"} {"id": "test_548", "sentence1": "The news interview setting revolves around sets of questions and answers-naively, one may assume the interviewer to be the sole questioner.", "sentence2": "media dialog has steadily deviated from this rigid structure, tending toward the broadly conversational (Fairclough, 1988).", "label": "contrasting"} {"id": "test_549", "sentence1": "Natural language inference (NLI) data has proven useful in benchmarking and, especially, as pretraining data for tasks requiring language understanding.", "sentence2": "the crowdsourcing protocol that was used to collect this data has known issues and was not explicitly optimized for either of these purposes, so it is likely far from ideal.", "label": "contrasting"} {"id": "test_550", "sentence1": "Longer texts offer the potential for discourse-level inferences, the addition of which should yield a dataset that is more difficult, more diverse, and less likely to contain trivial artifacts.", "sentence2": "one might expect that asking annotators to read full paragraphs should increase the time required to create a single example; time which could potentially be better spent creating more examples.", "label": "contrasting"} {"id": "test_551", "sentence1": "Giving annotators difficult and varying constraints could encourage creativity and prevent annotators from falling into patterns in their writing that lead to easier or more repetitive data.", "sentence2": "as with the use of longer contexts in PARAGRAPH, this protocol risks substantially slowing the annotation process.", "label": "contrasting"} {"id": "test_552", "sentence1": "Our chief results on transfer learning are conclusively negative: All four interventions yield substantially worse transfer performance than our base MNLI data collection protocol.", "sentence2": "we also observe promising signs that all four of our interventions help to reduce the prevalence of artifacts in the generated hypotheses that reveal the label.", "label": "contrasting"} {"id": "test_553", "sentence1": "We bring up textual entailment as a unified solver for such NLP problems.", "sentence2": "current research of textual entailment has not spilled much ink on the following questions: (i) How well does a pretrained textual entailment system generalize across domains with only a handful of domainspecific examples?", "label": "contrasting"} {"id": "test_554", "sentence1": "Thus, various stress-testing datasets have been proposed that probe NLI models for simple lexical inferences (Glockner et al., 2018), quantifiers (Geiger et al., 2018), numerical reasoning, antonymy and negation (Naik et al., 2018).", "sentence2": "despite the heavy usage of conjunctions in English, there is no specific NLI dataset that tests their understanding in detail.", "label": "contrasting"} {"id": "test_555", "sentence1": "We presented some initial solutions via adversarial training and a predicate-aware RoBERTa model, and achieved some reasonable performance gains on CONJNLI.", "sentence2": "we also show limitations of our proposed methods, thereby encouraging future work on CONJNLI for better understanding of conjunctive semantics.", "label": "contrasting"} {"id": "test_556", "sentence1": "This method is later extended to a hierarchical setting with a pre-defined hierarchy (Meng et al., 2019); ConWea (Mekala and Shang, 2020) leverages contextualized representation techniques to provide contextualized weak supervision for text classification.", "sentence2": "all these techniques consider only the text data and don't leverage metadata information for classification.", "label": "contrasting"} {"id": "test_557", "sentence1": "Note that the same hypothesis can also be made at the paragraph level.", "sentence2": "a major limitation of this approach is that paragraph sizes vary widely, ranging from a single word to a considerably huge block of text.", "label": "contrasting"} {"id": "test_558", "sentence1": "We use bidirectional contextual representation (Devlin et al., 2018) for encoding article text.", "sentence2": "contrary to document representation using BERT (Adhikari et al., 2019), which is not adequate for large text documents, we first segment articles organically based on sections.", "label": "contrasting"} {"id": "test_559", "sentence1": "Finally, there has been some work on directly training a model to extract entities and associated negation constraints (Bhatia et al., 2019).", "sentence2": "these works usually assume the availability of good quality annotated negated entities.", "label": "contrasting"} {"id": "test_560", "sentence1": "This behavior unique to S C+R is safe for the noisy data filtering task since it can successfully detect lower-quality pairs with high precision.", "sentence2": "improperly underestimating some acceptable pairs (i.e., low recall) is one downside of S C+R , and we discuss its influences in Section 6.3.", "label": "contrasting"} {"id": "test_561", "sentence1": "TXtract (Karamanolakis et al., 2020) incorporated the categorical structure into the value tagging system.", "sentence2": "these methods suffer from irrelevant articles and is not able to filter out noisy answers.", "label": "contrasting"} {"id": "test_562", "sentence1": "Similar to prior work (Lee et al., 2017), our training objective is to maximize the probability of the correct antecedent (cluster) for each mention span.", "sentence2": "rather than considering all correct antecedents, we are only interested in the cluster for the most recent one.", "label": "contrasting"} {"id": "test_563", "sentence1": "These plots show that models have relatively modest memory usage during inference.", "sentence2": "their usage grows in training, due to gradients and optimizer parameters.", "label": "contrasting"} {"id": "test_564", "sentence1": "This would \"correct\" the training objective to match prior work.", "sentence2": "this did not have a noticeable effect on performance.", "label": "contrasting"} {"id": "test_565", "sentence1": "Likewise, we were able to train a competitive model for which only the SpanBERT encoder from Joshi et al. (2019) was retained and the span scorer and pairwise scorer were randomly initialized.", "sentence2": "we opted not to use that for the full experiments because training was more expensive in time.", "label": "contrasting"} {"id": "test_566", "sentence1": "Neural generation models based on different strategies like softtemplate (Wiseman et al., 2018; Ye et al., 2020), copy-mechanism (See et al., 2017), content planning (Reed et al., 2018; Moryossef et al., 2019), and structure awareness Colin and Gardent, 2019) have achieved impressive results.", "sentence2": "existing studies are primarily focused on fully supervised setting requiring substantial labeled annotated data for each subtask, which restricts their adoption in real-world applications.", "label": "contrasting"} {"id": "test_567", "sentence1": "The work closest to our concept is Switch-GPT-2 (Chen et al., 2020b), which fits the pre-trained GPT-2 model as the decoder part to perform table-to-text generation.", "sentence2": "their knowledge encoder is still trained from scratch, which compromises the performance.", "label": "contrasting"} {"id": "test_568", "sentence1": "GPT (Radford, 2018) and GPT-2 (Radford et al., 2019) use a leftto-right Transformer decoder to generate a text sequence token-by-token, which lacks an encoder to condition generation on context.", "sentence2": "MASS (Song et al., 2019) and BART (Lewis et al., 2019) both employ a Transformer-based encoder\u0002decoder framework, with a bidirectional encoder over corrupted (masked) text and a left-to-right decoder reconstructing the original text.", "label": "contrasting"} {"id": "test_569", "sentence1": "However, these autoencoding methods are not applicable to text generation where bidirectional contexts are not available.", "sentence2": "an autoregressive model, such as GPT (Radford, 2018; Radford et al., 2019), is only trained to encode unidirectional context (either forward or backward).", "label": "contrasting"} {"id": "test_570", "sentence1": "A difference between UniLMs and PALM is that UniLMs are not fully autoregressive in the pre-training process.", "sentence2": "pALM reduces the mismatch between pre-training and context-conditioned generation tasks by forcing the decoder to predict the continuation of text input on an unlabeled corpus.", "label": "contrasting"} {"id": "test_571", "sentence1": "For instance in Figure 1, players might prefer the command \"move rug\" over \"knock on door\" since the door is nailed shut.", "sentence2": "even the state-of-the-art game-playing agents do not incorporate such priors, and instead rely on rule-based heuristics (Hausknecht et al., 2019a) or handicaps provided by the learning environment (Hausknecht et al., 2019a;Ammanabrolu and Hausknecht, 2020) to circumvent these issues.", "label": "contrasting"} {"id": "test_572", "sentence1": "In a slightly different setting, Urbanek et al. (2019) trained BERT (Devlin et al., 2018) to generate contextually relevant dialogue utterances and actions in fantasy settings.", "sentence2": "these approaches are game-specific and do not use any reinforcement learning to optimize gameplay.", "label": "contrasting"} {"id": "test_573", "sentence1": "When k is small, CALM (n-gram) benefits from its strong action assumption of one verb plus one object.", "sentence2": "this assumption also restricts CALM (n-gram) from generating more complex actions (e.g. \u2018open case with key\u2019) that CALM (GPT\u00022) can produce.", "label": "contrasting"} {"id": "test_574", "sentence1": "It is really the complex actions captured when k > 10 that makes GPT-2 much better than ngram.", "sentence2": "though k = 20, 30, 40 achieve similar overall performance, they achieve different results for different games.", "label": "contrasting"} {"id": "test_575", "sentence1": "It is interesting that CALM (w/ Jericho) is significantly better than CALM (GPT-2) on the games of Temple and Deephome (non-trivial scores achieved), which are not the games with ClubFloyd scripts added.", "sentence2": "games like 905 and moonlit have scripts added, but do not get improved.", "label": "contrasting"} {"id": "test_576", "sentence1": "Natural Language Generation (NLG) is a challenging problem in Natural Language Processing (NLP)-the complex nature of NLG tasks arise particularly in the output space.", "sentence2": "to text classification or regression problems with finite output space, generation could be seen as a combinatorial optimization problem, where we often have exponentially many options |V | (here |V | is the size of the vocabulary and is the sentence length).", "label": "contrasting"} {"id": "test_577", "sentence1": "It is possible to stop training the decomposition model based on downstream QA accuracy.", "sentence2": "training a QA model on each decom-position model checkpoint (1) is computationally expensive and (2) ties decompositions to a specific, downstream QA model.", "label": "contrasting"} {"id": "test_578", "sentence1": "The similarity between BERT sentence embeddings can be reduced to the similarity between BERT context embeddings h T c h c 2 .", "sentence2": "as shown in Equation 1, the pretraining of BERT does not explicitly involve the computation of h T c h c .", "label": "contrasting"} {"id": "test_579", "sentence1": "Note that BERT sentence embeddings are produced by averaging the context embeddings, which is a convexitypreserving operation.", "sentence2": "the holes violate the convexity of the embedding space.", "label": "contrasting"} {"id": "test_580", "sentence1": "We argue that NATSV can help eliminate anisotropy but it may also discard some useful information contained in the nulled vectors.", "sentence2": "our method directly learns an invertible mapping to isotropic latent space without discarding any information.", "label": "contrasting"} {"id": "test_581", "sentence1": "The transferred model both increased the quality and diversity of the generation.", "sentence2": "the transferred model exhibits narrower vocabulary usage.", "label": "contrasting"} {"id": "test_582", "sentence1": "Statistic-based automatic metrics, such as BLEU (Papineni et al., 2002), mostly rely on the degree of word overlap between a dialogue response and its corresponding gold response.", "sentence2": "due to the ignorance of the underlying semantic of a response, they are biased and correlate poorly with human judgements in terms of response coherence (Liu et al., 2016).", "label": "contrasting"} {"id": "test_583", "sentence1": "For example, BLEU computes the geometric average of the n-gram precisions.", "sentence2": "they can not cope with the one-to-many problem and have weak correlations with human judgements (Liu et al., 2016).", "label": "contrasting"} {"id": "test_584", "sentence1": "ADEM proposed by Lowe et al. (2017) achieves higher correlations with human judgements than the statistic-based metrics, which is trained with human-annotated data in a supervised manner.", "sentence2": "it is time-consuming and expensive to obtain large amounts of annotated data.", "label": "contrasting"} {"id": "test_585", "sentence1": "From the example in the first row, we can see that the score given by our metric is closer to the human score than the other two baseline metrics.", "sentence2": "in the second-row example, our metric performs poorly.", "label": "contrasting"} {"id": "test_586", "sentence1": "In this hard case, the topics of the model response are relevant to the dialogue context so that both our GRADE and BERT-RUBER, as learning-based metrics, deem that the response greatly matches the context.", "sentence2": "the truth is that the model response is more likely a response for the previous utterance U1 rather than U2, which is hard for metrics to recognize.", "label": "contrasting"} {"id": "test_587", "sentence1": "Such tasks include probing syntax (Hewitt and Manning, 2019; Lin et al., 2019; Tenney et al., 2019a), semantics (Yaghoobzadeh et al., 2019), discourse features (Chen et al., 2019; Liu et al., 2019; Tenney et al., 2019b), and commonsense knowledge (Petroni et al., 2019; Poerner et al., 2019).", "sentence2": "appropriate criteria for selecting a good probe is under debate.", "label": "contrasting"} {"id": "test_588", "sentence1": "BLEURT (Sellam et al., 2020) applies fine tuning of BERT, including training on prior human judgements.", "sentence2": "our work exploits parallel bitext and doesn't require training on human judgements.", "label": "contrasting"} {"id": "test_589", "sentence1": "We find that a copy of the input is almost as probable as beam search output for the Prism model.", "sentence2": "the model trained on ParaBank 2 prefers its own beam search output to a copy of the input.", "label": "contrasting"} {"id": "test_590", "sentence1": "We find that the probability of sys as estimated by an LM, as well as and the cosine distance between LASER embeddings of sys and ref, both have decent correlation with human judgments and are complementary.", "sentence2": "cosine distance between LASER embeddings of sys and src have only weak correlation.", "label": "contrasting"} {"id": "test_591", "sentence1": "Its creation is a manifestation of creativity, and, as such, hard to automate.", "sentence2": "since the development of creative machines is a crucial step towards real artificial intelligence, automatic poem generation is an important task at the intersection of computational creativity and natural language generation, and earliest attempts date back several decades; see Goncalo Oliveira (2017) for an overview.", "label": "contrasting"} {"id": "test_592", "sentence1": "Since those differ significantly in style from the poems in KnownTopicPoems and Unknown-TopicPoems, we do not train our language model directly on them.", "sentence2": "we make use of the fact that sonnets follow a known rhyming scheme, and leverage them to train a neural model to produce rhymes, which will be explained in detail in Subsection 3.2.", "label": "contrasting"} {"id": "test_593", "sentence1": "In particular, the generated poems seem to be more fluent and coherent than the alternatives.", "sentence2": "they do not relate to any specific topic, which probably causes the drop in quality for poeticness, where this model always performs worse than NeuralPoet.", "label": "contrasting"} {"id": "test_594", "sentence1": "The increase of K continues to bring benefits until K = 4.", "sentence2": "performance begins to drop when K > 3.", "label": "contrasting"} {"id": "test_595", "sentence1": "Neural Graph Encoding Graph Attention Networks (GAT) (Velickovic et al., 2018) incorporates attention mechanism in feature aggregation, RGCN (Schlichtkrull et al., 2018) proposes relational message passing which makes it applicable to multi-relational graphs.", "sentence2": "they only perform single-hop message passing and cannot be interpreted at path level.", "label": "contrasting"} {"id": "test_596", "sentence1": "RGCNs (Schlichtkrull et al., 2018) generalize GCNs by performing relationspecific aggregation, making it applicable to multirelational graphs.", "sentence2": "these models do not distinguish the importance of different neighbors or relation types and thus cannot provide explicit relational paths for model behavior interpretation.", "label": "contrasting"} {"id": "test_597", "sentence1": "This problem is clearly related a continual learning (CL) (Chen and Liu, 2018; Parisi et al., 2019; Li and Hoiem, 2017; Wu et al., 2018; Schwarz et al., 2018; Hu et al., 2019; Ahn et al., 2019), which also aims to learn a sequence of tasks incrementally.", "sentence2": "the main objective of the current CL techniques is to solve the catastrophic forgetting (CF) problem (McCloskey and Cohen, 1989).", "label": "contrasting"} {"id": "test_598", "sentence1": "That is, in learning each new task, the network parameters need to be modified in order to learn the new task.", "sentence2": "this modification can result in accuracy degradation for the previously learned tasks.", "label": "contrasting"} {"id": "test_599", "sentence1": "For sentiment classification, recent deep learning models have been shown to outperform traditional methods (Kim, 2014;Devlin et al., 2018;Shen et al., 2018;Qin et al., 2020).", "sentence2": "these models don't retain or transfer the knowledge to new tasks.", "label": "contrasting"} {"id": "test_600", "sentence1": "MTL is often considered the upper bound of continual learning because it trains all the tasks together.", "sentence2": "its loss is the sum of the losses of all tasks, which does not mean it optimizes for every individual task.", "label": "contrasting"} {"id": "test_601", "sentence1": "Automated radiology report generation has the potential to reduce the time clinicians spend manually reviewing radiographs and streamline clinical care.", "sentence2": "past work has shown that typical abstractive methods tend to produce fluent, but clinically incorrect radiology reports.", "label": "contrasting"} {"id": "test_602", "sentence1": "In this work we focused on developing abstractive techniques as was done by past work on the MIMIC-CXR dataset (Liu et al., 2019; Boag et al., 2019).", "sentence2": "in the future we intend to combine the abstractive methods developed in this work with retrieval methods to further improve upon our framework.", "label": "contrasting"} {"id": "test_603", "sentence1": "Sentence fusion has the lowest TER, indicating that obtaining the fused targets requires only a limited number of local edits.", "sentence2": "these edits require modeling the discourse relation between the two input sentences, since a common edit type is predicting the correct discourse connective (Geva et al., 2019).", "label": "contrasting"} {"id": "test_604", "sentence1": "Meng and Rumshisky (2018) propose a global context layer (GCL) to store/read the solved TLINK history upon a pre-trained pair-wise classifier.", "sentence2": "they find slow converge when training the GCL and pair-wise classifier simultaneously.", "label": "contrasting"} {"id": "test_605", "sentence1": "Meanwhile, a two-way deliberation decoder (Xia et al., 2017) was used for response generation.", "sentence2": "the relationship between the dialogue history and the last utterance is not well studied.", "label": "contrasting"} {"id": "test_606", "sentence1": "As a result, capturing the incongruity between modalities is significant for multi-modal sarcasm detection.", "sentence2": "the existing models for multi-modal sarcasm detection either concatenate the features from multi modalities (Schifanella et al., 2016) or fuse the information from different modalities in a designed manner (Cai et al., 2019).", "label": "contrasting"} {"id": "test_607", "sentence1": "Thus, an effective sarcasm detector is beneficial to applications like sentiment analysis, opinion mining (Pang and Lee, 2007), and other tasks that require people's real sentiment.", "sentence2": "the figurative nature of sarcasm makes it a challenging task (Liu, 2010).", "label": "contrasting"} {"id": "test_608", "sentence1": "For the store owner, the task of correctly identifying the buying-intent utterances is paramount.", "sentence2": "the number of utterances related to searching for products is expected to be significantly higher, thus biasing the classifier toward this intent.", "label": "contrasting"} {"id": "test_609", "sentence1": "Other approaches for data balancing can include weak-labeling of available unlabeled data (Ratner et al., 2020), or even active learning (Settles, 2009).", "sentence2": "both of these approaches require additional domain data which is not always available.", "label": "contrasting"} {"id": "test_610", "sentence1": "We focused our evaluation on the Semantic Utterance Classification (SUC) domain which is characterized by highly imbalanced data.", "sentence2": "it is desirable to validate the applicability of our general balancing approach on other textual domains.", "label": "contrasting"} {"id": "test_611", "sentence1": "Knowledge+BERT turns out to be the strongest baseline, outperforming the other three baselines, which also shows the importance of leveraging external knowledge for the OAC2 task.", "sentence2": "our model achieves superior performance over Knowledge+BERT which indicates leveraging domain-specific knowledge indeed helps.", "label": "contrasting"} {"id": "test_612", "sentence1": "This is expected as these tweets include less standard words, such as insults.", "sentence2": "except for perhaps emotion detection and offensive language identification, the difference is not significant, considering that the original RoBERTa tokenizer was not trained on Twitter text.", "label": "contrasting"} {"id": "test_613", "sentence1": "In normal attention, all n-grams are weighted globally and short n-grams may dominate the attention because they occur much more frequently than long ones and are intensively updated.", "sentence2": "there are cases that long n-grams can play an important role in parsing when they carry useful context and boundary information.", "label": "contrasting"} {"id": "test_614", "sentence1": "These results could be explained by that frequent short n-grams dominate the general attentions so that the long ones containing more contextual information fail to function well in filling the missing information in the span representation, and thus harm the understanding of long spans, which results in inferior results in complete match score.", "sentence2": "the categorical span attention is able to weight n-grams in different length separately, so that the attentions are not dominated by high-frequency short n-grams and thus reasonable weights can be assigned to long n-grams.", "label": "contrasting"} {"id": "test_615", "sentence1": "This is not surprising because short n-grams occur more frequently and are thus updated more times than long ones.", "sentence2": "the models with CATSA show a different weight distribution (the blue bars) among n-grams with different lengths, which indicates that the CATSA module could balance the weights distribution and thus enable the model to learn from infrequent long n-grams.", "label": "contrasting"} {"id": "test_616", "sentence1": "Since the distances between the boundary positions of the wrongly predicted spans (highlighted in red) are relatively long, the baseline system, which simply represents the span as subtraction of the hidden vectors at the boundary positions, may fail to capture the important context information within the text span.", "sentence2": "the span representations used in our model are enhanced by weighted n-gram information and thus contain more context information.", "label": "contrasting"} {"id": "test_617", "sentence1": "Because training with soft targets provides smoother output distribution, T/S learning could outperform the single model training (Li et al., 2014;Hinton et al., 2015;Meng et al., 2018).", "sentence2": "does a teacher always outperform a student?", "label": "contrasting"} {"id": "test_618", "sentence1": "All models benefit from the increasing of D as expected.", "sentence2": "it is clear that Recurrence Online is the best performing model when D is small.", "label": "contrasting"} {"id": "test_619", "sentence1": "We observe that the use of Static Rebalancing (Equation 6) instead, which is an extreme version of AISLe, is better than not resampling at all.", "sentence2": "it is unable to reach the performance of AISLe on coverage metrics.", "label": "contrasting"} {"id": "test_620", "sentence1": "Kreutzer and Sokolov (2018) proposed to jointly learn to segment and translate by using hierarchical RNN (Graves, 2016), but the method is not model-agnostic and slow due to the increased sequence length of characterlevel inputs.", "sentence2": "our method is model-agnostic and operates on the word-level.", "label": "contrasting"} {"id": "test_621", "sentence1": "Kudo (2018) also report scores using n-best decoding, which averages scores from n-best segmentation results.", "sentence2": "n-best decoding is n-times time consuming compared to the standard decoding method.", "label": "contrasting"} {"id": "test_622", "sentence1": "Starting from machine translation, it has been shown that subword regularization can improve the robustness of NLP models in various tasks (Kim, 2019;Provilkov et al., 2019;Drexler and Glass, 2019;M\u00fcller et al., 2019).", "sentence2": "subword regularization relies on the unigram language models to sample candidates, where the language models are optimized based on the corpus-level statistics from training data with no regard to the translation task objective.", "label": "contrasting"} {"id": "test_623", "sentence1": "SurfCon (Wang et al., 2019b) discovered synonyms on privacy-aware clinical data by utilizing the surface form information and the global context information.", "sentence2": "they suffer from either low precision or low recall.", "label": "contrasting"} {"id": "test_624", "sentence1": "In this example, the word victim in the first English sentence is identified by our tagger as a human entity.", "sentence2": "its French translation victime is feminine by definition, and cannot be assigned another gender regardless of the context, causing a false positive result.", "label": "contrasting"} {"id": "test_625", "sentence1": "Furthermore, they explain IBT as a way to better approximate the true posterior distribution with the target-to-source model.", "sentence2": "it is unclear how their heuristic objective relates to the ideal objective of maximizing the model's marginal likelihood of the target language monolingual data.", "label": "contrasting"} {"id": "test_626", "sentence1": "Deep neural models have demonstrated promising results in text classification tasks (Kim, 2014;Zhang et al., 2015;Howard and Ruder, 2018), owing to their strong expressive power and less requirement for feature engineering.", "sentence2": "the deeper and more complex the neural model, the more it is essential for them to be trained on substantial amount of training data.", "label": "contrasting"} {"id": "test_627", "sentence1": "Several natural language processing methods, including deep neural network (DNN) models, have been applied to address this problem.", "sentence2": "these methods were trained with hard-labeled data, which tend to become over-confident, leading to degradation of the model reliability.", "label": "contrasting"} {"id": "test_628", "sentence1": "In the training step, the model is trained to maximize the output probability of the correct class.", "sentence2": "some studies reported that the deep learning classifier trained with hard-labeled data (1 for correct class, 0 for else) tends to become over-confident (Nixon et al., 2019;Thulasidasan et al., 2019).", "label": "contrasting"} {"id": "test_629", "sentence1": "These deep learning methods can effectively generate abstractive document summaries by directly optimizing pre-defined goals.", "sentence2": "the meeting summarization task inherently bears a number of challenges that make it more difficult for end-to-end training than document summarization.", "label": "contrasting"} {"id": "test_630", "sentence1": "As shown in Figure 1, the medical report generation system should generate correct and concise reports for the input images.", "sentence2": "data imbalance may reduce the quality of automatically generated reports.", "label": "contrasting"} {"id": "test_631", "sentence1": "To improve the clinical correctness of the generated reports, Liu et al. (2019a) and Irvin et al. (2019) adopted clinically coherent rewards for RL with CheXpert Labeler (Irvin et al., 2019), a rule-based finding mention annotator", "sentence2": "in the medical domain, no such annotator is available in most cases other than English chest X-ray reports.", "label": "contrasting"} {"id": "test_632", "sentence1": "The input data, which comprise a set of finding labels, can be augmented easily by adding or removing a finding label automatically.", "sentence2": "the augmentation cost is higher for the target reports than the input data because the target reports are written in natural language.", "label": "contrasting"} {"id": "test_633", "sentence1": "In addition to (c) above, we apply the modification process to the finding labels predicted by the image diagnosis module.", "sentence2": "it is too expensive to evaluate the model in this condition because the cost of radiologist services is too high.", "label": "contrasting"} {"id": "test_634", "sentence1": "For comparison with the previous image captioning approaches, we used BLEU-1, BLEU-2, BLEU-3, and BLEU-4 metrics calculated by the nlg-eval 10 library.", "sentence2": "word-overlap based metrics, such as BLEU, fail to assume the factual correctness of generated reports.", "label": "contrasting"} {"id": "test_635", "sentence1": "In ERNIE, entity embeddings are learned by TransE (Bordes et al., 2013), which is a popular transitionbased method for knowledge representation learning (KRL).", "sentence2": "transE cannot deal with the modeling of complex relations , such as 1-to-n, n-to-1 and n-to-n relations.", "label": "contrasting"} {"id": "test_636", "sentence1": "An intuitive way for vanilla GCN to exploit these labels is to encode different types of dependency relation with different convolutional filters, which is similar to RGCN (Kipf and Welling, 2017).", "sentence2": "rGCN suffers from over-parameterization, where the number of parameters grows rapidly with the number of relations.", "label": "contrasting"} {"id": "test_637", "sentence1": "We considered each sentence as a single claim to keep our experimental setting clean and avoid noise from an automatic claim extractor.", "sentence2": "some generations contain multiple claims that could be independently assessed.", "label": "contrasting"} {"id": "test_638", "sentence1": "By thorough error analysis, we realize that for the order h-t-r (t-h-r follows the same logic), the model has to predict all t with regard to h in the second time step, without constraints from the r, and this makes every possible entity to be a prediction candidate.", "sentence2": "the model is unable to eliminate no-relation entity pairs at the third time step, thus the model is prone to feed entity pairs to the classification layer with an low odds (low recall) but high confidence (high precision).", "label": "contrasting"} {"id": "test_639", "sentence1": "Over the past two decades, significant progress has been made in the development of word embedding techniques (Lund and Burgess, 1996; Bengio et al., 2003; Bullinaria and Levy, 2007; Mikolov et al., 2013b; Pennington et al., 2014).", "sentence2": "existing word embedding methods do not handle numerals adequately and cannot directly encode the numeracy and magnitude of a numeral Naik et al. (2019).", "label": "contrasting"} {"id": "test_640", "sentence1": "These studies conclude that when certain nouns are dropped from the dominant language modality, multimodal models are capable of properly using the semantics provided by the image.", "sentence2": "unlike this work, their explorations are limited to nouns and not expanded to other types of words.", "label": "contrasting"} {"id": "test_641", "sentence1": "They demonstrate high compression rate with little loss of performance.", "sentence2": "they compress only the input embedding and not the softmax layer for language modeling and machine translation.", "label": "contrasting"} {"id": "test_642", "sentence1": "Specifically, they regard each segmentation criterion as a single task under the framework of multi-task learning, where a shared layer is used to extract the criteriainvariant features, and a private layer is used to extract the criteria-specific features.", "sentence2": "it is unnecessary to use a specific private layer for each criterion.", "label": "contrasting"} {"id": "test_643", "sentence1": "As shown in several previous research (Pang et al., 2016;Yang et al., 2016;Mitra et al., 2017;Xiong et al., 2017;Devlin et al., 2018), interaction-focused models usually achieve better performances for text pair tasks.", "sentence2": "it is difficult to serve these types of models for applications involving large inference sets in practice.", "label": "contrasting"} {"id": "test_644", "sentence1": "However, it is difficult to serve these types of models for applications involving large inference sets in practice.", "sentence2": "text embeddings from dual encoder models can be learned independently and thus pre-computed, leading to faster inference efficiency but at the cost of reduced quality.", "label": "contrasting"} {"id": "test_645", "sentence1": "Recently the PreTTR model (MacAvaney et al., 2020) aimed to reduce the query-time latency of deep transformer networks by pre-computing part of the document term representations.", "sentence2": "their model still required modeling the full document/query input length in the head, thus limiting inference speedup.", "label": "contrasting"} {"id": "test_646", "sentence1": "One of the most effective ways to reduce the running time is to reduce the input sequence length.", "sentence2": "as Table 4 reveals, blindly truncating the input to a BERT model will lead to a quick performance drop.", "label": "contrasting"} {"id": "test_647", "sentence1": "When the head is transformerbased, the two-stage training plays an important role: the AUC ROC improves from 0.891 to 0.930.", "sentence2": "the gain introduced by using two-stage training is less significant in other approaches such as DE-FFNN and DIPAIRFFNN.", "label": "contrasting"} {"id": "test_648", "sentence1": "Recall that, in our framework, each encoder outputs its first few token embeddings as the input to the head, and we end to end to train the model to force the encoder to push the information of the input text into those outputted embeddings.", "sentence2": "it is unclear to us what those outputted embeddings actually learn.", "label": "contrasting"} {"id": "test_649", "sentence1": "We formalize word reordering as a combinatorial optimization problem to find the permutation with the highest probability estimated by a POS-based language model.", "sentence2": "it is computationally difficult to obtain the optimal word order.", "label": "contrasting"} {"id": "test_650", "sentence1": "With some classifiers, we reached the same F1-score as when training on the original dataset, which is 20x larger.", "sentence2": "performance varied markedly between classifiers.", "label": "contrasting"} {"id": "test_651", "sentence1": "This demonstrates that GPT-2 significantly increased the vocabulary range of the training set, specifically with offensive words likely to be relevant for toxic language classification.", "sentence2": "there is a risk that human annotators might not label GPT-2-generated documents as toxic.", "label": "contrasting"} {"id": "test_652", "sentence1": "TABLE-BERT is a BERT-base model that similar to our approach directly predicts the truth value of the statement.", "sentence2": "the model does not use special embeddings to encode the table structure but relies on a template approach to format the table as natural language.", "label": "contrasting"} {"id": "test_653", "sentence1": "In language tasks, adversarial training brings wordlevel robustness by adding input noise, which is beneficial for text classification.", "sentence2": "it lacks sufficient contextual information enhancement and thus is less useful for sequence labelling tasks such as chunking and named entity recognition (NER).", "label": "contrasting"} {"id": "test_654", "sentence1": "Masked language model (Devlin et al., 2018) smooths this inconsistency by applying replacement of tokens for some data while masking the rest (equivalent to word dropout).", "sentence2": "the replacement in masked language model is randomly chosen from the full vocabulary, but the substitution in real sce\u0002narios follows some distribution (e.g. replacing \u201cMassachusetts\u201d with a location name is more likely than an animal name), which is not considered in masked language model.", "label": "contrasting"} {"id": "test_655", "sentence1": "Zhao et al. (2018) proposed the learning scheme to generate a gender-neutral version of Glove, called GN-Glove, which forces preserving the gender information in pre-specified embedding dimensions while other embedding dimensions are inferred to be gender-neutral.", "sentence2": "learning new word embeddings for large-scale corpus can be difficult and expensive.", "label": "contrasting"} {"id": "test_656", "sentence1": "Table 4 shows that there are constant performance degradation effects for all debiasing methods from the original embedding.", "sentence2": "our methods minimized the degradation of performances across the baseline models.", "label": "contrasting"} {"id": "test_657", "sentence1": "A clear-cut solution to this problem is to focus more on samples that are more relevant to the target task during pretraining.", "sentence2": "this requires a task-specific pretraining, which in most cases is computational or time prohibitive.", "label": "contrasting"} {"id": "test_658", "sentence1": "Further, Moreo et al. (2019) concatenates label embedding with word embeddings.", "sentence2": "this approach cannot be directly implemented into PLMs since the new (concatenated) embedding is not compatible with the pretrained parameters.", "label": "contrasting"} {"id": "test_659", "sentence1": "The growing interest in argument mining and computational argumentation brings with it a plethora of Natural Language Understanding (NLU) tasks and corresponding datasets.", "sentence2": "as with many other NLU tasks, the dominant language is English, with resources in other languages being few and far between.", "label": "contrasting"} {"id": "test_660", "sentence1": "The ZS and TT baselines are almost always outperformed by the best translate-train model.", "sentence2": "when a large-scale English corpus is available (Figure 2b), the TT baseline becomes comparable to the best translate-train models.", "label": "contrasting"} {"id": "test_661", "sentence1": "Dialogue policy learning for Task-oriented Dialogue Systems (TDSs) has enjoyed great progress recently mostly through employing Reinforcement Learning (RL) methods.", "sentence2": "these approaches have become very sophisticated.", "label": "contrasting"} {"id": "test_662", "sentence1": "With respect to success rate, DiaAdv manages to achieve the highest performance by 6% compared to the second highest method GDPL.", "sentence2": "diaAdv is not able to beat GdPL in terms of average turns.", "label": "contrasting"} {"id": "test_663", "sentence1": "As to DiaSeq, it can achieve almost the same performance as GDPL from different perspectives while GDPL has a slightly higher F1 score.", "sentence2": "the potential cost benefits of DiaSeq are huge since it does not require a user simulator in the training loop.", "label": "contrasting"} {"id": "test_664", "sentence1": "Beyond this, DiaMultiClass does not benefit from the increase in expert dialogues and starts to fluctuate between 55% and 59%.", "sentence2": "di-aSeq can achieve higher performance when there are only 10% expert dialogue pairs and the success rate increases with the number of available expert dialogues.", "label": "contrasting"} {"id": "test_665", "sentence1": "The proposed methods can achieve state-of-the-art performance suggested by existing approaches based on Reinforcement Learning (RL) and adversarial learning.", "sentence2": "we have demonstrated that our methods require fewer training efforts, namely the domain knowledge needed to design a user simulator and the intractable parameter tuning for RL or adversarial learning.", "label": "contrasting"} {"id": "test_666", "sentence1": "Our evaluation settings hiding one of the three inputs to the MCQA models -are similar to Kaushik and Lipton 2018's partial input settings which were designed to point out the existence of dataset artifacts in reading comprehension datasets.", "sentence2": "we argue that our results additionally point to a need for more robust training methodologies and propose an improved training approach.", "label": "contrasting"} {"id": "test_667", "sentence1": "Among these, hyperedge replacement grammar (HRG) has been explored for parsing into semantic graphs (Habel, 1992;Chiang et al., 2013).", "sentence2": "parsing with HRGs is not practical due to its complexity and large number of possible derivations per graph (Groschwitz et al., 2015).", "label": "contrasting"} {"id": "test_668", "sentence1": "Drawing on this result, a recent work by Fancellu et al. (2019) introduces recurrent neural network RDGs, a sequential decoder that models graph generation as a rewriting process with an underlying RDG.", "sentence2": "despite the promising framework the approach in FA19 2 falls short in several aspects.", "label": "contrasting"} {"id": "test_669", "sentence1": "Composition is constrained by the rank of a nonterminal so to ensure that at each decoding step the model is always aware of the placement of reentrant nodes.", "sentence2": "we do not ensure semantic well-formedness in that words are predicted separately from their fragments and we do not rely on alignment information.", "label": "contrasting"} {"id": "test_670", "sentence1": "Early CLWE approaches required expensive parallel data (Klementiev et al., 2012; T\u00e4ckstr\u00f6m et al., 2012).", "sentence2": "later approaches rely on high-coverage bilingual dictionaries (Gliozzo and Strapparava, 2006; Faruqui and Dyer, 2014; or smaller \"seed\" dictionaries (Gouws and S\u00f8gaard, 2015; Artetxe et al., 2017).", "label": "contrasting"} {"id": "test_671", "sentence1": "Multilingual BERT performs well on zero-shot cross-lingual transfer (Wu and Dredze, 2019; Pires et al., 2019) and its performance can be further improved by considering target-language documents through self-training (Dong and de Melo, 2019).", "sentence2": "our approach does not require multilingual language models and sometimes outperforms multilingual BERT using a monolingual BERT student.", "label": "contrasting"} {"id": "test_672", "sentence1": "Neural network-based models augmented with unsupervised pre-trained knowledge have achieved impressive performance on text summarization.", "sentence2": "most existing evaluation methods are limited to an in-domain setting, where summarizers are trained and evaluated on the same dataset.", "label": "contrasting"} {"id": "test_673", "sentence1": "Bigpatent B also exhibits relatively higher copy rate in summary but the copy segments is shorter than CNNDM.", "sentence2": "bigaptent b, Xsum obtain higher sentence fusion score, which suggests that the proportion of fused sentences in these two datasets are high.", "label": "contrasting"} {"id": "test_674", "sentence1": "2) BART (SOTA system) is superior over other abstractive models and even comparable with extractive models in terms of stiffness (ROUGE).", "sentence2": "it is robust when transferring between datasets as it possesses high stableness (ROUGE).", "label": "contrasting"} {"id": "test_675", "sentence1": "Typical sources of transfer loss concern differences in features between domains (Blitzer et al., 2007;Ben-David et al., 2010).", "sentence2": "other factors may govern model degradation for depression clas-sification.", "label": "contrasting"} {"id": "test_676", "sentence1": "Topical nuances in language may appropriately reflect elements of identity associated with mental health disorders (i.e. traumatic experiences, coping mechanisms)", "sentence2": "if not contextualized during model training, this type of signal has the potential to raise several false alarms upon application to new populations.", "label": "contrasting"} {"id": "test_677", "sentence1": "Figure 3b shows that BiLSTM uses 35% of context for short sentences, 20% for medium, and only 10% for long sentences.", "sentence2": "bERT leverages fixed 75% of context words regardless of the sentence length.", "label": "contrasting"} {"id": "test_678", "sentence1": "VQA requires techniques from both image recognition and natural language processing, and most existing works use Convolutional Neural Networks (CNNs) to extract visual features from images and Recurrent Neural Networks (RNNs) to generate textual features from questions, and then combine them to generate the final answers.", "sentence2": "most existing VQA datasets are created in a way that is not suitable as training data for real-world applications.", "label": "contrasting"} {"id": "test_679", "sentence1": "VQA models following this setting take characteristics of all answer candidates like word embeddings as the input to make a selection (Sha et al., 2018; Jabri et al., 2016).", "sentence2": "in the open-ended setting, there is neither prior knowledge nor answer candidates provided, and the model can respond with any freeform answers.", "label": "contrasting"} {"id": "test_680", "sentence1": "We conjecture that this is because when we have limited amount of target data, having more prior knowledge is beneficial to model performance, while having more target data will make prior knowledge less helpful.", "sentence2": "our method can stably improve the performance because it sufficiently makes use of target data and source data.", "label": "contrasting"} {"id": "test_681", "sentence1": "Importantly, we do not define a precondition event as an absolute requirement for the target (the door opening) to occur in all scenarios.", "sentence2": "we do require that the target event likely would not have occurred in the current context.", "label": "contrasting"} {"id": "test_682", "sentence1": "This reveals the source of improvement in attack success rate between GENETICATTACK and TEXTFOOLER to be more lenient constraint application.", "sentence2": "gE-NETICATTACK's genetic algorithm is far more computationally expensive, requiring over 40x more model queries.", "label": "contrasting"} {"id": "test_683", "sentence1": "Gilmer et al. (2018) laid out a set of potential constraints for the attack space when generating adversarial examples, which are each useful in different real-world scenarios.", "sentence2": "they did not discuss NLP attacks in particular.", "label": "contrasting"} {"id": "test_684", "sentence1": "Object manipulation and configuration is another subject that has been studied along with language and vision grounding (Bisk et al., 2016;Wang et al., 2016;Li et al., 2016;Bisk et al., 2018).", "sentence2": "most studies focus on addressing the problem in relatively simple environments from a third-person view.", "label": "contrasting"} {"id": "test_685", "sentence1": "It is possible for instructions to be written that can pass all automated checks and still be of poor quality.", "sentence2": "there is no quick and reliable way to automatically check if an instruction passes the tests but is still vague or misleading.", "label": "contrasting"} {"id": "test_686", "sentence1": "There are several benchmarks (Wen et al., 2017;El Asri et al., 2017;Eric and Manning, 2017;Wei et al., 2018) to evaluate the performance of neural models for goal-oriented dialog.", "sentence2": "these benchmarks assume a world of a \"perfect\" user who always provides precise, con- cise, and correct utterances.", "label": "contrasting"} {"id": "test_687", "sentence1": "Zhao and Eskenazi (2018) created SimDial, which simulates spoken language phenomena, e.g. self-repair and hesitation. Sankar et al. (2019) introduce utterance-level and word\u0002level perturbations on various benchmarks.", "sentence2": "such variations have been largely artificial and do not reflect the \"natural variation\" commonly found in naturally occuring conversational data.", "label": "contrasting"} {"id": "test_688", "sentence1": "The conversational activity patterns (denoted by A) handle the main business of conversation, i.e. the user request and the services provided by agent.", "sentence2": "conversation management patterns help the user and agent to manage the conversation itself.", "label": "contrasting"} {"id": "test_689", "sentence1": "Since the models are evaluated only on the agent responses present in the original test set, additional user and agent utterances for incorporating natural variation do not affect performance Table 5: Ablation results for GLMP model on SMD too much.", "sentence2": "sMD is a real-world dataset of human-to-human conversations collected by crowdsourcing and we observe a much higher drop across both BLEU and Ent F1 scores.", "label": "contrasting"} {"id": "test_690", "sentence1": "This resulted in some novelty in the data collected and prevented the user utterances to be repetitive.", "sentence2": "to control data collection, the participants were asked to follow a set of instructions which resulted in user utterances largely focused on the task.", "label": "contrasting"} {"id": "test_691", "sentence1": "The dataset is the largest currently as it has largest context complexity and state complexity (based on all possible combinations of customer and agent context features, like number of flights in the database, number of airlines, airport codes and dialogue action states), in comparison to other existing datasets mentioned above.", "sentence2": "the authors don't share details on how the dataset was collected and instructions provided to the participants", "label": "contrasting"} {"id": "test_692", "sentence1": "As shown in Figure 2, a correlation exists between some relevant events (such as the first joint press release) and the number of articles published.", "sentence2": "a higher volume of articles does not always correlate with higher disagreement rates between annotators: interestingly, it seems that some events (such as the merger agreement) spread more uncertainty around the merger than others (such as the start of the antitrust trial).", "label": "contrasting"} {"id": "test_693", "sentence1": "Similar to our baselines ScRNN (Sakaguchi et al., 2017) and MUDE (Wang et al., 2019), Li et al. (2018) proposed a nested RNN to hierarchically encode characters to word representations, then correct each word using a nested GRU .", "sentence2": "these previous works either only train models on natural misspellings (Sakaguchi et al., 2017) or synthetic misspellings , and only focus on denoising the input texts from orthographic perspective without leveraging the retained semantics of the noisy input.", "label": "contrasting"} {"id": "test_694", "sentence1": "These LMs captures the probability of a word or a sentence given their context, which plays a crucial role in correcting real-word misspellings.", "sentence2": "all of the LMs mentioned are based on subword embeddings, such as WordPiece (Peters et al., 2018) or Byte Pair Encoding (Gage, 1994) to avoid OOV words.", "label": "contrasting"} {"id": "test_695", "sentence1": "XLNet (Yang et al., 2019) also marginalize over all possible factorizations.", "sentence2": "their work is focused on the conditional distribution p(y|x), and they do not marginalize over all possible factorizations of the joint distribution.", "label": "contrasting"} {"id": "test_696", "sentence1": "KERMIT is a generative joint distribution model that also learns all possible factorizations.", "sentence2": "kERMIT is constrained to two languages, while MGLM is a generative joint distribution model across any/all languages/text while learning all possible factorizations of the joint distribution.", "label": "contrasting"} {"id": "test_697", "sentence1": "Multilingual Neural Language Model (Wada and Iwata, 2018) uses a shared encoder and language-dependent decoders to generate word embeddings and evaluate word alignment tasks.", "sentence2": "our work unifies the neural architecture with a straightforward stack of self-attention layers.", "label": "contrasting"} {"id": "test_698", "sentence1": "Our work focused on a specific instantiation of channels as languages.", "sentence2": "mGLm is not limited to only languages and can generalize to other notions of channels.", "label": "contrasting"} {"id": "test_699", "sentence1": "suggest a misleading \"PERSON\" label 19 because of their context features, so that an incorrect NER prediction is expected if treating the three types of syntactic information equally.", "sentence2": "the syntactic constituents give strong indication of the correct label through the word \"Rights\" for a \"LAW\" entity.", "label": "contrasting"} {"id": "test_700", "sentence1": "Recently, neural models play dominant roles in NER because of their effectiveness in capturing contextual information in the text without requir\u0002ing to extract manually crafted features (Huang et al., 2015; Lample et al., 2016; Strubell et al., 2017; Zhang and Yang, 2018; Peters et al., 2018; Yadav and Bethard, 2018; Cetoli et al., 2018; Ak\u0002bik et al., 2018, 2019; Chen et al., 2019; Devlin et al., 2019; Zhu and Wang, 2019; Liu et al., 2019b; Baevski et al., 2019; Yan et al., 2019; Xu et al., 2019a; Zhu et al., 2020; Luo et al.).", "sentence2": "to enhance NER, it is straightforward to incorporate more knowledge to it than only modeling from contexts.", "label": "contrasting"} {"id": "test_701", "sentence1": "It is thus far more important to evaluate various seed configurations that various target documents.", "sentence2": "we wanted to keep the computational cost of evaluation reasonably small, so either the number of seed configurations had to be reduced or the number of target documents for each configuration.", "label": "contrasting"} {"id": "test_702", "sentence1": "In a recent edition, Rabelo et al. (2019) used a BERT model fine-tuned on a provided training set in a supervised manner, and achieved the highest F-score among all teams.", "sentence2": "due to the reasons discussed in Section 4, their approach is not consistent with the nearest neighbor search, which is what we are aiming for.", "label": "contrasting"} {"id": "test_703", "sentence1": "Deep neural models have achieved impressive success in many areas.", "sentence2": "their interpretability and explainability have remained broadly limited", "label": "contrasting"} {"id": "test_704", "sentence1": "Such methods extract parts of the model input that are important to the output according to some criterion.", "sentence2": "they are not suited to evaluate NL explanations that are not part of the input, which motivates our new simulatability metric.", "label": "contrasting"} {"id": "test_705", "sentence1": "Among others, dependency trees help to directly link the aspect term to the syntactically related words in the sentence, thus facilitating the graph convolutional neural networks (GCN) (Kipf and Welling, 2017) to enrich the representation vectors for the aspect terms.", "sentence2": "there are at least two major issues in these graph-based models that should be addressed to boost the performance.", "label": "contrasting"} {"id": "test_706", "sentence1": "These statements represent generic commonsense hypotheses about social behaviors and their acceptability that are held as norms in a society.", "sentence2": "such normative judgments can also be strengthened or weakened given appropriate context.", "label": "contrasting"} {"id": "test_707", "sentence1": "Hyperbolic spaces offer a mathematically appealing approach for learning hierarchical representations of symbolic data.", "sentence2": "it is not clear how to integrate hyperbolic components into downstream tasks.", "label": "contrasting"} {"id": "test_708", "sentence1": "Many models that fuse visual and linguistic features have been proposed.", "sentence2": "few models consider the fusion of linguistic features with multiple visual features with different sizes of receptive fields, though the proper size of the receptive field of visual features intuitively varies depending on expressions.", "label": "contrasting"} {"id": "test_709", "sentence1": "Zhao et al. (2018) also proposes a model with a structure that fuses multiple scales and languages for weakly supervised learning.", "sentence2": "they use concatenation as the method of fusion, whereas we use FiLM.", "label": "contrasting"} {"id": "test_710", "sentence1": "The paraphrase ratio of the augmented training set remains similar as the original set.", "sentence2": "the ratio increases in the augmented testing set indicating the paraphrase clusters are sparser in the testing set.", "label": "contrasting"} {"id": "test_711", "sentence1": "Another line of work learns Hornclause style reasoning rules from the KG and stores them in its parameters (Rocktaschel and Riedel, 2017; Das et al., 2018; Minervini et al., 2020).", "sentence2": "these parametric approaches work with a fixed set of entities and it is unclear how these models will adapt to new entities.", "label": "contrasting"} {"id": "test_712", "sentence1": "For example, the performance (MRR) of ROTATE model (Sun et al., 2019) drops by 11 points (absolute) on WN18RR in this setting (\u00a73.4).", "sentence2": "we show that with new data, the performance of our model is consistent as it is able to seamlessly reason with the newly arrived data.", "label": "contrasting"} {"id": "test_713", "sentence1": "Recent works (Teru et al., 2020;Wang et al., 2020) learn entity independent relation representations and hence allow them to handle unseen entities.", "sentence2": "they do not perform contextual reasoning by gathering reasoning paths from similar entities.", "label": "contrasting"} {"id": "test_714", "sentence1": "As a result, many of the query relations were different from what was present in the splits of NELL-995 and hence is not a good representative.", "sentence2": "we report test results for the best hyper-parameter values that we got on this validation set.", "label": "contrasting"} {"id": "test_715", "sentence1": "For sentences in GENIA, the number of candidate regions generated by HiRe is 77.9% less than that of the enumeration method discarding 1.3% long entities and more than that of (Zheng et al., 2019).", "sentence2": "the true recall of candidate regions generated by the enumeration method and HiRe are 98.7% and 98.1%, respectively.", "label": "contrasting"} {"id": "test_716", "sentence1": "HiRe without HRR employs Average Word Representation (denoted as AWR) instead with precision 78.3%, recall 73.7% and F1 measure 75.9%.", "sentence2": "to HiRe AWR , the absolute F1 measure improvement of HiRe HRR is 0.6%.", "label": "contrasting"} {"id": "test_717", "sentence1": "Therefore some recent researches attempt to endow the bots with proactivity through external knowledge to transform the role from a listener to a speaker with a hypothesis that the speaker expresses more just like a knowledge disseminator.", "sentence2": "along with the proactive manner introduced into a dialogue agent, an issue arises that, with too many knowledge facts to express, the agent starts to talks endlessly, and even completely ignores what the other expresses in dialogue sometimes, which greatly harms the interest of the other chatter to continue the conversation.", "label": "contrasting"} {"id": "test_718", "sentence1": "Models facilitated with external knowledge indeed generate more meaningful responses than peers that train only on the source-target dialogue dataset.", "sentence2": "these models tend to fall into another situation where the machine agent talks too much ignoring what the other has said, let alone the inappropriate use of knowledge.", "label": "contrasting"} {"id": "test_719", "sentence1": "What's more, with copy mechanism, CopyNet, DeepCopy, and Initiative-Imitate perform better in terms of fluency and coherence because of the utilizing of proper knowledge.", "sentence2": "comparing CopyNet and DeepCopy with Seq2Seq attn , the Engagement becomes worse because too much knowledge harms the ability to react to the proposed ques-tion very likely.", "label": "contrasting"} {"id": "test_720", "sentence1": "Lexical resources such as WordNet (Miller, 1995) capture such synonyms (say, tell) and hypernyms (whisper, talk), as well as antonyms, which can be used to refer to the same event when the arguments are reversed ([a] 0 beat [a] 1 , [a] 1 lose to [a] 0 ).", "sentence2": "WordNet\u2019s coverage is insufficient, in particular, missing contextspecific paraphrases (e.g. (hide, launder), in the context of money).", "label": "contrasting"} {"id": "test_721", "sentence1": "During training, each encoder learns a language model specific to an individual MT source, yielding diversity among experts in the final system.", "sentence2": "in order to improve robustness of each encoder to translation variability, inputs to each encoder are shuffled by some tuned probability p shuffle .", "label": "contrasting"} {"id": "test_722", "sentence1": "Previous LSTM-based ensemble approaches propose training full parallel networks and ensemble at the final decoding step.", "sentence2": "we found this was too expensive given the nonrecurrent Transformer model.", "label": "contrasting"} {"id": "test_723", "sentence1": "Most current multi-hop relation reasoning models require a good amount of training data (fact triples) for each query relation.", "sentence2": "the relation frequency distribution in KB is usually longtail , showing that a large portion of relations only have few-shot fact triples for model training.", "label": "contrasting"} {"id": "test_724", "sentence1": "We look into the task of generalizing word embeddings: extrapolating a set of pre-trained word embeddings to words out of its fixed vocabulary, without extra access to contextual information (e.g. example sentences or text corpus).", "sentence2": "the more common task of learning word embeddings, or often just word embedding, is to obtain distributed representations of words directly from large unlabeled text.", "label": "contrasting"} {"id": "test_725", "sentence1": "We omit the prediction time for KVQ-FH, as we found it hard to separate the actual inference time from time used for other processes such as batching and data transfer between CPU and GPU.", "sentence2": "we believe the overall trend should be similar as for the training time.", "label": "contrasting"} {"id": "test_726", "sentence1": "In this field, the supervised methods, ranging from the conventional graph models (McCallum et al., 2000; Malouf, 2002; McCallum and Li, 2003; Settles, 2004) to the dominant deep neural methods (Collobert et al., 2011; Huang et al., 2015; Lample et al., 2016; Gridach, 2017; Liu et al., 2018; Zhang and Yang, 2018; Jiang et al., 2019; Gui et al., 2019), have achieved great success.", "sentence2": "these supervised methods usually require large scale labeled data to achieve good performance, while the annotation of NER data is often laborious and time-consuming.", "label": "contrasting"} {"id": "test_727", "sentence1": "Then, it finetunes the model pretrained on the source task (with the output layer being replaced) using the re-annotated data to perform the target task.", "sentence2": "it is worth noting that the NER labels of words are contextdependent.", "label": "contrasting"} {"id": "test_728", "sentence1": "However, given that style transfer can be viewed as a monolingual machine translation (MT) task, and that seq2seq models such as the transformer have shown to outperform unsupervised methods in multi-lingual MT when a sufficiently large parallel corpus is available (Lample et al., 2018; Artetxe et al., 2019; Subramanian et al., 2018), in our opinion it is expected that seq2seq would outperform unsupervised approaches if parallel data is available for style transfer.", "sentence2": "to the best of our knowledge, a parallel corpus for style transfer currently does not exist.", "label": "contrasting"} {"id": "test_729", "sentence1": "Therefore, finding the value for STAcc is trivial once C has been found.", "sentence2": "finding a value for C is the main issue for the metric, since it depends on evaluating the set of generated outputs how many of them were converted successfully.", "label": "contrasting"} {"id": "test_730", "sentence1": "We show that in a data-rich setting, with sufficient training examples, our approach outperforms a classification-based encoder-only model.", "sentence2": "our sequence-to-sequence model appears to be far more data-efficient, significantly outperforming BERT with few training examples in a data-poor setting.", "label": "contrasting"} {"id": "test_731", "sentence1": "We discuss this question in Section 5.4.", "sentence2": "as a preview, we find that the choice of target tokens has a large impact on effectiveness in some circumstances, and these experiments shed light on why T5 works well for document ranking.", "label": "contrasting"} {"id": "test_732", "sentence1": "While the approach can exploit pretrained knowledge when fine-tuning the latent representations, the final mapping (i.e., the fully-connected layer) needs to be learned from scratch (since it is randomly initialized).", "sentence2": "T5 can exploit both pretrained knowledge and knowledge gleaned from fine-tuning in learning task-specific latent representations as well as the mapping to relevance decisions; specifically, we note that T5 is pretrained with tasks whose outputs are \"true\" and \"false\".", "label": "contrasting"} {"id": "test_733", "sentence1": "It has long been observed that most relation tuples follow syntactic regularity, and many syntactic patterns have been designed for extracting tuples, such as TEXTRUNNER (Banko et al., 2007) and ReVerb (Fader et al., 2011).", "sentence2": "it is difficult to design high coverage syntactic patterns, although many extensions have been proposed, such as WOE (Wu and Weld, 2010), OLLIE (Mausam et al., 2012), ClausIE (Corro and Gemulla, 2013), Standford Open IE , PropS and OpenIE4 (Mausam, 2016).", "label": "contrasting"} {"id": "test_734", "sentence1": "If BERT considers either name to be a common French name, then a correct answer is insufficient evidence for factual knowledge about the entity Jean Marais.", "sentence2": "if neither Jean nor Marais are considered French, but a correct answer is given regardless, we consider it sufficient evidence of factual knowledge.", "label": "contrasting"} {"id": "test_735", "sentence1": "Existing approaches to improve generalization in QA either are only applicable when there exist multiple training domains (Talmor and Berant, 2019;Takahashi et al., 2019; or rely on models and ensembles with larger capacity (Longpre et al., 2019;Su et al., 2019;.", "sentence2": "our novel debiasing approach can be applied to both single and multi-domain scenarios, and it improves the model generalization without requiring larger pre-trained language models.", "label": "contrasting"} {"id": "test_736", "sentence1": "Mahabadi et al. (2020) handle multiple biases jointly and show that their debiasing methods can improve the performance across datasets if they fine-tune their debiasing methods on each target dataset to adjust the debiasing parameters.", "sentence2": "the impact of their method is unclear on generalization to unseen evaluation sets.", "label": "contrasting"} {"id": "test_737", "sentence1": "In addition, some works (Sukhbaatar et al., 2015;Madotto et al., 2018;Wu et al., 2019) have considered integrating KBs in a task-oriented dialogue system to generate a suitable response and have achieved promising performance.", "sentence2": "these methods either are limited by predefined configurations or do not scale to large KBs.", "label": "contrasting"} {"id": "test_738", "sentence1": "However, as the KBs continue to grow in the real-world scenarios, such end-to-end methods of directly encoding and integrating whole KBs will eventually result in inefficiency and incorrect responses.", "sentence2": "some works may put the user utterances through a semantic parser to obtain executable logical forms and apply this symbolic query to the KB to retrieve entries based on their attributes.", "label": "contrasting"} {"id": "test_739", "sentence1": "In the transformer, the representation of each query token gets updated by self-attending to the representations of all the query tokens and graph nodes in the previous layer.", "sentence2": "the representation of each graph node gets updated by self-attending only to its graph neighbors according to the connections of the sparsely connected transformer as well as all query tokens.", "label": "contrasting"} {"id": "test_740", "sentence1": "Thus, we explore the possibility of augmenting the user-generated data with synthetic data in order to train a better model.", "sentence2": "one needs to be careful with data augmentation using synthetic data as it inevitably has a different distribution.", "label": "contrasting"} {"id": "test_741", "sentence1": "However, when the size of task-related data is large enough, using a pre-trained model does not deliver much benefits (TS2 and TS all ft).", "sentence2": "finetuning clearly improves the performance of TS2 pt in both human and automatic metrics, where using only 1k domain data already produces satisfying scores in human metrics.", "label": "contrasting"} {"id": "test_742", "sentence1": "Multi2 OIE yields the highest recall for all languages by approximately 20%p.", "sentence2": "argOE has relatively high precision, but low recall negatively impacts its F1 score.", "label": "contrasting"} {"id": "test_743", "sentence1": "Most of the systems available for French fit in those three approaches.", "sentence2": "none of these systems have been thoroughly compared to each other, even with the release, in 2013, of a large coreference annotated corpus (Muzerelle et al., 2014), since each system uses slightly different versions of the corpus for evaluation (e.g. different train and test sets).", "label": "contrasting"} {"id": "test_744", "sentence1": "CROC uses a feature indicating whether a mention is a new entity in the text, i.e. whether a mention is the first in its coreference chain.", "sentence2": "this feature is usually only available when the corpus has been previously annotated or after the coreference resolution task: this is why we removed it.", "label": "contrasting"} {"id": "test_745", "sentence1": "It is difficult to study coreference errors of an end-to-end system, since it is not possible to fully separate mention misidentifications from coreference issues.", "sentence2": "it allows a better understanding of error source.", "label": "contrasting"} {"id": "test_746", "sentence1": "One reason for this is that the newspapers and magazines that were used for our corpus tend to contain quite complex texts (political commentary and reports in historical German).", "sentence2": "we also observed some systematic difficulties to apply our annotation system that is rooted in narrative theory to journalistic writing, e.g.", "label": "contrasting"} {"id": "test_747", "sentence1": "We generally follow this idea: reported is more summarizing and less precise, while indirect ST&WR can usually be read as a transformation of direct ST&WR that allows us to reconstruct the 'original' quote in more detail.", "sentence2": "there are sentences that follow the typical structure of indirect ST&WRa framing clause and a dependent subordinate clause containing the content -but do not allow such a reconstruction", "label": "contrasting"} {"id": "test_748", "sentence1": "There is a large number of tools and software packages providing access to data repositories such as NLTK (Loper and Bird, 2002) or Spacy 1 .", "sentence2": "many of these resources are not powerful enough to exploit this data to their full extent.", "label": "contrasting"} {"id": "test_749", "sentence1": "Therefore it is normally only available through a web GUI hosted by the Institute of the Czech National Corpus.", "sentence2": "we obtained a tabular text file with the Czech-German alignment on request.", "label": "contrasting"} {"id": "test_750", "sentence1": "Reddit users make less use of argumentation proposition types in general: they use less normative language than the candidates and express less desire than Republican candidates.", "sentence2": "they use reported speech often, partly because their discussions occurred after the debates had occurred.", "label": "contrasting"} {"id": "test_751", "sentence1": "We found that models trained through multi-task learning where the primary task consists of argument component classification and the secondary task consists of specificity classification almost always outperform models that only perform argument component classification.", "sentence2": "the corpus used in our previous study is not publicly available and therefore our previous results are not reproducible by other members of the research community.", "label": "contrasting"} {"id": "test_752", "sentence1": "In our prior work (Lugini and Litman, 2018) on argument component classification for discussions, we used oversam\u0002pling to alleviate the class imbalance present in argumenta\u0002tion labels (which is also present in the Discussion Tracker corpus).", "sentence2": "since our Discussion Tracker experiments also include 3 task multi-task learning, oversampling with respect to argumentation labels might have negative impact on other tasks.", "label": "contrasting"} {"id": "test_753", "sentence1": "These hypotheses are motivated by our observation of differences between collaboration label distributions across argumentative moves.", "sentence2": "given the different unit of analysis for the annotation of collaboration (turn) versus argumentation and specificity (argument discourse unit), for the multi-task learning setting the collaboration annotations have been converted to BIO format in order to have one annotation per argument move 2 .", "label": "contrasting"} {"id": "test_754", "sentence1": "Punctuation symbols often indicates segment boundaries.", "sentence2": "there may be cases where EDUs are not segmented.", "label": "contrasting"} {"id": "test_755", "sentence1": "The baseline model fails to identify the comparative \u2018enough . . . to' as a correlative and does not segment the sentence.", "sentence2": "training for syntactic features allowed the model to correctly identify this construct and hence perform correct segmentation.", "label": "contrasting"} {"id": "test_756", "sentence1": "As suspected, the baseline model performs poorly when the sentences are longer.", "sentence2": "formulating the problem in an alternate fashion and injecting syntax make the model perform much better.", "label": "contrasting"} {"id": "test_757", "sentence1": "Despite its simple mechanism, this algorithm comes with a high bias, which is unfavorable for learning new directions within the data.", "sentence2": "multi-View-Training (Zhou and Goldman, 2004;S\u00f8gaard, 2010) tries to compensate this bias by different views of the data.", "label": "contrasting"} {"id": "test_758", "sentence1": "They collected explicit argument pairs with freely omissible discourse connectives which can be dropped independently of the context without changing the interpretation of the discourse relation.", "sentence2": "sporleder and Lascarides (2008) argued training on explicit argument pairs was not a good strategy.", "label": "contrasting"} {"id": "test_759", "sentence1": "Note that Wu et al. (2017) collected explicit argument pairs using a similar method.", "sentence2": "they only used argument pairs located within the same sentence while we do not apply this constraint.", "label": "contrasting"} {"id": "test_760", "sentence1": "We also use a content extraction tool to extract article content from an HTML file, and apply a shingling-based method to identify near-duplicate articles.", "sentence2": "our systems differ in two major ways.", "label": "contrasting"} {"id": "test_761", "sentence1": "Furthermore, some datasets are available for evaluating additional dimensions of essay quality in English (Mathias and Bhattacharyya, 2018).", "sentence2": "only a few evaluation datasets are available for Japanese writings, and even fewer Japanese learner essay datasets are.", "label": "contrasting"} {"id": "test_762", "sentence1": "We created the feature-based models using linguistic features based on (Lee and Hasebe, 2017).", "sentence2": "whether these features are enough to perform AES is unclear.", "label": "contrasting"} {"id": "test_763", "sentence1": "As a result, the neural approach for AES has been actively studied in recent years (Taghipour and Ng, 2016).", "sentence2": "no neural-network-based AES system is available for the Japanese language; furthermore, the BERT model has not been applied for an AES task with multiple dimensions thus far.", "label": "contrasting"} {"id": "test_764", "sentence1": "The feature-based model predicted a score that was two points lower than the actual content and organization trait scores in essay A and a score that was two points lower than the organization trait score in essay B.", "sentence2": "the neural-network-based model predicted the score correctly for essay A and predicted a score that was only one point lower than the actual language trait score in essay B.", "label": "contrasting"} {"id": "test_765", "sentence1": "Further, this model may provide a high score for an unexpected input.", "sentence2": "the neural-network-based model predicted low scores for all columns.", "label": "contrasting"} {"id": "test_766", "sentence1": "Reported results were obtained with traditional machine learning methods and, to some extent, it would be interesting to test more recent classification methods, such as deep neural networks.", "sentence2": "the corpus might not be large enough for such an approach, which further motivates this kind of experiment.", "label": "contrasting"} {"id": "test_767", "sentence1": "They proved to be a strong baseline in the binary classification task, outperforming the surface text-based features and the graph-based deep semantic features.", "sentence2": "on the five-level classification task, they were outperformed by all other feature sets.", "label": "contrasting"} {"id": "test_768", "sentence1": "To create such collections, there is a substantial need for automatic approaches that can distinguish the documents of interest for a collection out of the large collections (of millions in size) from Web Archiving institutions.", "sentence2": "the patterns of the documents of interest can differ substantially from one document to another, which makes the automatic classification task very challenging.", "label": "contrasting"} {"id": "test_769", "sentence1": "The time-domain features offer a simple way to analyse audio signals and are directly extracted from the samples of the audio signal (waveform).", "sentence2": "frequencydomain features are extracted from the sound spectrum, a representation of the distribution of the frequency content of sounds (Giannakopoulos and Pikrakis, 2014d).", "label": "contrasting"} {"id": "test_770", "sentence1": "The corpus contains tweets annotated with 28 emotions categories and captures the language used to express an emotion explicitly and implicitly.", "sentence2": "the availability of datasets created specifically for languages other than English is very limited.", "label": "contrasting"} {"id": "test_771", "sentence1": "For several years, affect in speech has been encoded using discrete categories such as anger, sadness or neutral speech.", "sentence2": "in many recent papers, researchers preferred using affective dimensions.", "label": "contrasting"} {"id": "test_772", "sentence1": "Regarding the effect of hesitation on fundamental frequency (f0), a study on German spontaneous speech (Mixdorff and Pfitzinger, 2005) found no impact of hesitations marked by fillers on the overall f0 pattern at the utterance level.", "sentence2": "a study relying on synthesized speech (Carlson et al., 2006) in Swedish showed a moderate effect of the f0 slope on perceived hesitation, as well as a moderate effect of the insertion of creaky voice.", "label": "contrasting"} {"id": "test_773", "sentence1": "Thus all adaptations of the speaking styles to different degrees of hesitation are individual as well and cannot be summed up as a group mean.", "sentence2": "the tendencies of the individual changes remain similar across the group.", "label": "contrasting"} {"id": "test_774", "sentence1": "Child language studies are crucial in improving our understanding of child well-being; especially in determining the factors that impact happiness, the sources of anxiety, techniques of emotion regulation, and the mechanisms to cope with stress.", "sentence2": "much of this research is stymied by the lack of availability of large child-written texts.", "label": "contrasting"} {"id": "test_775", "sentence1": "Valence was significantly negatively associated with arousal (r = -.06, p < .001), although the effect was small, suggesting minimal collinearity.", "sentence2": "correlations with dominance (both A-D and D-V) were much stronger and significant (p < .001).", "label": "contrasting"} {"id": "test_776", "sentence1": "Interpreting it at face value, we might conclude that the results reflect increased capabilities in emotion regulation (i.e., being more in control of one's emotions) (Zimmermann and Iwanski, 2014).", "sentence2": "we are hesitant to make this conclusion because individual words likely have poor correspondence with emotion regulation, which involves complex processes.", "label": "contrasting"} {"id": "test_777", "sentence1": "In our results, we observe that when maximum of the 3 CCCs computed on each pair is low, the predicted satisfaction is likely to be bad.", "sentence2": "if this maximum is high, the predicted satisfaction is likely to be good.", "label": "contrasting"} {"id": "test_778", "sentence1": "It often carries both positive and negative feelings.", "sentence2": "since this label is quite infrequent, and not available in all subsets of the data, we annotated it with an additional Beauty/Joy or Sadness label to ensure annotation consistency.", "label": "contrasting"} {"id": "test_779", "sentence1": "The results of the crowdsourcing experiment, on the other hand, are a mixed bag as evidenced by a much sparser distribution of emotion labels.", "sentence2": "we note that these differences can be caused by 1) the disparate training procedure for the experts and crowds, and 2) the lack of opportunities for close supervision and on-going training of the crowds, as opposed to the in-house expert annotators.", "label": "contrasting"} {"id": "test_780", "sentence1": "Nostalgia is still available in the gold standard (then with a second label Beauty/Joy or Sadness to keep consistency).", "sentence2": "confusion, Boredom and Other are not available in any sub-corpus.", "label": "contrasting"} {"id": "test_781", "sentence1": "This line of work is predominantly based on word-level supervision.", "sentence2": "we learn word ratings from document-level ratings.", "label": "contrasting"} {"id": "test_782", "sentence1": "Mean Star Rating, Binary Star Rating, and Regressions Weights learn exclusively from the available document-level gold data.", "sentence2": "one of the major advantages of the MLFFN is that it builds on pre-trained word embeddings, thus implicitly leveraging vast amounts of unlabeled text data.", "label": "contrasting"} {"id": "test_783", "sentence1": "Another seemingly obvious evaluation strategy would be to predict document-level ratings from derived word-level lexica using the empathic reactions dataset in a cross-validation setup.", "sentence2": "we found that this approach has two major drawbacks.", "label": "contrasting"} {"id": "test_784", "sentence1": "The clusters tend to be consistent regarding NE types, as illustrated in Table 3.", "sentence2": "both false positives (FP; non-NE entries in NE clusters) as well as false negatives (FN; NE entries in non-NE clusters) do occur.", "label": "contrasting"} {"id": "test_785", "sentence1": "The transcription of the production as well as the target form are generally available in an orthographic form.", "sentence2": "if we are interested primarily in oral production and as this oral production is sometimes restricted to isolated words, it is ultimately more important to place the analysis at the phonological level.", "label": "contrasting"} {"id": "test_786", "sentence1": "In this characterization, /p/ (+ coronal, + anterior) differs from /t/ (+ coronal) by one feature and similarly from /k/ (-anterior).", "sentence2": "/t/ (+ coronal, + anterior) differs from /k/ (-coronal, -anterior) by two features, which is not very satisfactory from an articulatory point of view where it would seem logical to respect the order /ptk/, that is to say /t/ equidistant from /p/ and /k/, /p/ and /k/ being more distant.", "label": "contrasting"} {"id": "test_787", "sentence1": "The maximum precision score that Terrier reached is 0.92 indicating that the optimum of 1.0 was never achieved.", "sentence2": "sODA reached the maximum score of 1.0 in five cases.", "label": "contrasting"} {"id": "test_788", "sentence1": "Indeed lemmatization, stemming, numeral masking and entity masking did not improve results.", "sentence2": "stop word filtering produces better word embeddings for this task.", "label": "contrasting"} {"id": "test_789", "sentence1": "Currently, the serial corpora provided by NLPCC are the mainstream evaluation benchmarks for Chinese EL.", "sentence2": "all of them stem from Chinese microblogs, which can be fairly short and noisy.", "label": "contrasting"} {"id": "test_790", "sentence1": "Evidently, mentions in the Hard document are rather ambiguous, as Hinrich, Chandler can refer to many different entities.", "sentence2": "easy document contains very obvious mentions such as BRICS, UN and the country names.", "label": "contrasting"} {"id": "test_791", "sentence1": "These comments are useful for the learner.", "sentence2": "comments are noise for an evaluation dataset because automatic evaluation methods utilizing corrected sentences typically rely on the matching rate between the system output and the corrected sentences to calculate a score.", "label": "contrasting"} {"id": "test_792", "sentence1": "In this model, we tokenized the learner sentence at the character level.", "sentence2": "we tokenized a corrected sentence at the word level.", "label": "contrasting"} {"id": "test_793", "sentence1": "Therefore, it turns out that a CNN-based method is effective for errors that can be corrected with only the local context (Chollampatt and Ng, 2018).", "sentence2": "both the NMT and SMT systems could hardly correct errors that needed to be considered in context, for example, abbreviation or formal and casual style errors.", "label": "contrasting"} {"id": "test_794", "sentence1": "The number of true positives (TP) in the NMT system was larger than that in the SMT system.", "sentence2": "the number of false positives (FP) in the NMT system was considerably larger than that in the SMT system.", "label": "contrasting"} {"id": "test_795", "sentence1": "Lang-8's original annotation contains annotator's comments that are noise for evaluation.", "sentence2": "our evaluation corpus does not contain such comments.", "label": "contrasting"} {"id": "test_796", "sentence1": "The second sentence does match the index query, so the full traversal is performed.", "sentence2": "because there is an intervening xcomp relation, the traversal fails.", "label": "contrasting"} {"id": "test_797", "sentence1": "In the above example, \u201c\u0906\u0924\u0902\u0915\u0940 \u0939\u092e\u0932\u093e\u201d (terrorist attack) is a multiword event trigger with annotation labels B_Event and I_Event respectively.", "sentence2": " \u2018\u0939\u092e\u0932\u093e' (attack) itself is an event trigger with annotation label B_Event.", "label": "contrasting"} {"id": "test_798", "sentence1": "The recognition of medical concepts and their attributes in EEG reports is vital for many applications requiring data-driven representation of EEG-specific knowledge, including decision support systems.", "sentence2": "the identification of the medical concepts in the EEG reports is not sufficient, as these concepts also exhibit clinically-relevant relations between them.", "label": "contrasting"} {"id": "test_799", "sentence1": "Since the Morphology best defines the EEG activities, we decided to use it as an anchor for each mention of an EEG activity in the EEG report.", "sentence2": "Morphology represents the type or \"form\" of an EEG activity, which may have multiple values, as seen in Table 2, therefore the Morphology remains also as an attribute of the EEG activities.", "label": "contrasting"} {"id": "test_800", "sentence1": "In deciding the nodes of the HAD, we have consulted the Epilepsy Syndrome and Seizure Ontology (ESSO) 2 , which encodes 2,705 classes with an upper ontology targeting epilepsy and selected the concepts that best describe EEG activities.", "sentence2": "eeG events, which are frequently mentioned in eeG reports as well, can be recognized only by identifying the text span where they are mentioned and their polarity and modality attributes.", "label": "contrasting"} {"id": "test_801", "sentence1": "In a controlled laboratory environment, participants used the Crowdee platform for performing the summary quality evaluation task.", "sentence2": "to the crowdsourcing study, all the participants were also instructed in a written form following the standard practice for laboratory tests.", "label": "contrasting"} {"id": "test_802", "sentence1": "Automatic analysis of connected speech by natural language processing techniques is a promising direction for diagnosing cognitive impairments.", "sentence2": "some difficulties still remain: the time required for manual narrative transcription and the decision on how transcripts should be divided into sentences for successful application of parsers used in metrics, such as Idea Density, to analyze the transcripts.", "label": "contrasting"} {"id": "test_803", "sentence1": "Prosodic features have been shown to be very effective to discriminate between different types of sentence boundaries and in general their usage reflects better results (Shriberg et al., 2009;Huang et al., 2014;Khomitsevich et al., 2015).", "sentence2": "to put prosodic features into practice we need alignments between the audio and its transcription, which is hard to obtain mainly due to the low quality of the recordings.", "label": "contrasting"} {"id": "test_804", "sentence1": "From these results, we can assume that sequenced-figures narratives bring linguistic features also present in retellings, but the reverse direction is not true, as we can see in Table 5.", "sentence2": "if a researcher will only work on retelling tasks, Table 5 shows that using only retelling datasets for training led to better results for the retelling task.", "label": "contrasting"} {"id": "test_805", "sentence1": "For a script that is not phonetic, e.g., Chinese characters, grapheme-tophoneme conversion is considered compulsory.", "sentence2": "as Hangul is phonetic, in other words, text in Hangul sounds as it is written, we stick with graphemes rather than converting them into phonemes.", "label": "contrasting"} {"id": "test_806", "sentence1": "Much of the information about temples is available as text in the open web, which can be utilized to conduct such a study.", "sentence2": "this information is not in the form of a learning resource, which can be readily used for such studies.", "label": "contrasting"} {"id": "test_807", "sentence1": "On the one hand high quality resources are needed that contain (English) glosses, part of speech tagging as well as underlying morphophonemes forms.", "sentence2": "the resource needs to be large enough for the wanted forms to occur in the data.", "label": "contrasting"} {"id": "test_808", "sentence1": "On a positive side, 50% of the sentences were different from one seed strategy to the other, suggesting for an approach where strategies are mixed.", "sentence2": "we also observed that (a) tends to yield more similar queries over time and (c) is too time-consuming for practical use.", "label": "contrasting"} {"id": "test_809", "sentence1": "When English is the source language, and Japanese is the target language, there are only 5 pairs in the test data where the source and target words are identical, i.e., cases where the copy baseline is correct.", "sentence2": "in the case of English being the target language, and Japanese being the source language, there are 270 pairs where the source and target words are identical.", "label": "contrasting"} {"id": "test_810", "sentence1": "Wikipedia is usually used as a high-quality freely available multilingual corpus as compared to noisier data such as Common Crawl.", "sentence2": "for the two languages under study, Wikipedia resulted to have too much noise: interference from other languages, text clearly written by non-native speakers, lack of diacritics and mixture of dialects.", "label": "contrasting"} {"id": "test_811", "sentence1": "For example, a sentence describing the English cricket team's victory over India could invoke negative sentiment, given the annotator's strong support of the latter.", "sentence2": "the actual label of such a statement would be positive because of the author's intention.", "label": "contrasting"} {"id": "test_812", "sentence1": "Logistic Regression offers marginally better performance than Linear-SVM in terms of precision (Precision for LR is 0.675).", "sentence2": "the former fails to outperform the latter in the other three metrics of evaluation.", "label": "contrasting"} {"id": "test_813", "sentence1": "The choice of approach has a fundamental effect on the end result: in the case of expansion (translation), the new wordnet will be fully meaning-aligned with the source language (English), which is ideal for cross-lingual uses: as most wordnets are already aligned with PWN, we get bilingual translations to all those languages 'for free'.", "sentence2": "a certain linguistic bias is introduced by the fact that only meanings for which English lexicalisations exist will appear in the wordnet.", "label": "contrasting"} {"id": "test_814", "sentence1": "A common form of sarcasm consists of a positive sentiment contrasted with a negative situation (Riloff et al., 2013), therefore it was likely that learning the emotional information of a text would facilitate the task of irony/sarcasm prediction.", "sentence2": "it was seen that the majority of Persian Twitter users include either humor, irony or sarcasm in their posts.", "label": "contrasting"} {"id": "test_815", "sentence1": "pre-trained a neural network model to predict emojis in the text and then transferred the model for different related tasks including sarcasm detection (Felbo et al., 2017).", "sentence2": "with approaches that use feature engineering to extract features , in (Amir et al., 2016) features are automatically extracted by learning user embeddings which requires users' preceding messages.", "label": "contrasting"} {"id": "test_816", "sentence1": "Please note that we carry out our analysis on the whole corpus.", "sentence2": " if one were interested only in the most reliable portions of the corpus, i.e. the cases all annotators agreed upon, different confidence thresholds can be set, as shown in Table 2.", "label": "contrasting"} {"id": "test_817", "sentence1": "The high frequency of idioms in persuasive and rhetorical language corroborates the statements by McCarthy (1998) that idioms are used for commenting on the world, rather than describing it, and Minugh 2008, who finds that idioms are used most often by those with some authority, especially when conveying 'received wisdom'.", "sentence2": "this genre distinction is still quite crude.", "label": "contrasting"} {"id": "test_818", "sentence1": "Due to the fact that the majority of indigenous languages were traditionally exclusively oral cultures, manuscripts typically do not play a primary role in the study of those languages.", "sentence2": "in particular handwritten notes that were created by researchers during fieldwork often are important information sources that, beyond other things, contain highly relevant information, ranging from object language data with attached translations and glossings, over lexical and grammatical descriptions to complex metadata, figural data and of course individual interpretation by the respective researcher.", "label": "contrasting"} {"id": "test_819", "sentence1": "The resulting derived resource on the one hand shows which data from the resource catalogue and which sessions from the corpora have been published in certain bibliographic items.", "sentence2": "it demonstrates that some have actually been published in different bibliographic items.", "label": "contrasting"} {"id": "test_820", "sentence1": "A popular and widely method is crawling different web pages.", "sentence2": "this is only possible under the assumption that there are sufficient web sites written in the target language.", "label": "contrasting"} {"id": "test_821", "sentence1": "In some cases, the large number of hapax is related to a poor quality of the corpus that might be caused by spelling errors or the presence of foreign words (Nagata et al., 2018).", "sentence2": "our scenario is expected given the agglutinative nature of the four target languages, so they might present a vast vocabulary diversity.", "label": "contrasting"} {"id": "test_822", "sentence1": "Crowdsourcing platforms, such as Amazon Mechanical Turk, have been an effective method for collecting such large amounts of data.", "sentence2": "difficulties arise when task-based dialogues require expert domain knowledge or rapid access to domain-relevant information, such as databases for tourism.", "label": "contrasting"} {"id": "test_823", "sentence1": "Prior crowdsourced wizarded data collections have divided the dialogue up into turns and each worker's job consists of one turn utterance generation given a static dialogue context, as in the MultiWoZ dataset (Budzianowski et al., 2018).", "sentence2": "this can limit naturalness of the dialogues by restricting forward planning, collaboration and use of memory that humans use for complex multi-stage tasks in a shared dynamic environment/context.", "label": "contrasting"} {"id": "test_824", "sentence1": "The first WordNet for Bulgarian was built in the BalkaNet project (Koeva and Genov, 2004).", "sentence2": "to this date the lexicon is not freely available.", "label": "contrasting"} {"id": "test_825", "sentence1": ". A free core-WordNet for Bulgarian was made available in the BulTreeBank Wordnet (Simov and Osenova, 2010), but unfortunately its size is rather small - 8 936 senses.", "sentence2": "it has very good quality, so when we had to choose a translation for a word in English, we first looked for a corresponding synset in the BulTreeBank Wordnet.", "label": "contrasting"} {"id": "test_826", "sentence1": "The dictionary is compatible with GF, and contains morphology and English-Bulgarian translations.", "sentence2": "the translations are not sense annotated.", "label": "contrasting"} {"id": "test_827", "sentence1": "Sense annotations are available only for the words exemplified with that sentence.", "sentence2": "in order to provide good translations, we sense tagged all words in the corpus.", "label": "contrasting"} {"id": "test_828", "sentence1": "The target audience of the digital literature on this platform is young people (roughly below 30 years of age), it thus has the potential of being biased in age.", "sentence2": "this is the age group that is the most fluent in and most likely to use written Cantonese and therefore it reflects the current usage of written Cantonese in the society.", "label": "contrasting"} {"id": "test_829", "sentence1": "Authors like Brysbaert and New (2009) suggest that a size of about 15 millions tokens guarantees a robust estimation of the term frequencies (i.e. which correlates well with psycholinguistic measures)", "sentence2": "because HKC is not particularly well resourced and that corpus data (especially spoken) is not always freely available, we made do with what is available at the time, and have much less than that target size.", "label": "contrasting"} {"id": "test_830", "sentence1": "In order to handle these cases (which become more frequent as the resources to be modelled become more 'scholarly') we decided to make etymology a class, Etymology, in Part 3.", "sentence2": "since etymologies usually represent an ordering of etymons (and, of course, etymons can be associated with more than one etymology and even more than one etymology for the same entry), we opted to create indirect rather than direct associations between etymologies and etymons.", "label": "contrasting"} {"id": "test_831", "sentence1": "If a lemma differ from one source to another, we create multiple entries and disambiguate them manually based on the definitions obtained from other resources.", "sentence2": "if the lemma is the same we fuse their information.", "label": "contrasting"} {"id": "test_832", "sentence1": "We need to enrich it and validate it by Old French or diachrony specialists.", "sentence2": "the manual process is long and tedious especially when it comes to enrich the lexicon by decreasing the silence rate.", "label": "contrasting"} {"id": "test_833", "sentence1": "In recent years, people have started investigating neologisms computationally (e.g. Ahmad (2000; Kerremans et al. (2011)), and online dictionaries and datasets provide convenient electronic versions of a word's year of first use.", "sentence2": "these resources vary in the amount of information they provide and are often limited to a handful of languages.", "label": "contrasting"} {"id": "test_834", "sentence1": "The post-editing speed here was lower, around 1.5K words/hour.", "sentence2": "the proportion of tags edited, 1.8%, is only slightly higher.", "label": "contrasting"} {"id": "test_835", "sentence1": "When given the same set of essays to evaluate and enough graded samples, AES systems tend to achieve high agreement levels with trained human raters (Taghipour and Ng, 2016).", "sentence2": "there is a sizeable literature in cognitive science, psychology and other social studies offering evidence that biases can create situations that lead us to make decisions that project our experiences and values onto others (Baron, 2007).", "label": "contrasting"} {"id": "test_836", "sentence1": "Here, the target language expression is typically either a translation or a definition.", "sentence2": "these cannot be reliably distinguished in an automated way, so that target language information is best represented as a definition rather than as a translation.", "label": "contrasting"} {"id": "test_837", "sentence1": "There are many existing Arabic corpora (Atwell, 2019).", "sentence2": "we are only interested in those that include Hadith or classical Arabic text in general.", "label": "contrasting"} {"id": "test_838", "sentence1": "In another study, a survey was conducted to enumerate the freely available Arabic corpora and stated the existence of one Hadith corpus.", "sentence2": "it was not accessible, mentioned or used in the literature (Zaghouani, 2017).", "label": "contrasting"} {"id": "test_839", "sentence1": "The number of tokens in the English Hadiths is larger than the Arabic version.", "sentence2": "the Arabic Hadiths are richer in vocabulary as it contains more unique words than the English version as shown in Table 4.", "label": "contrasting"} {"id": "test_840", "sentence1": "Applications that allow users to interact with technology via spoken or written natural language are emerging in all areas, and access to language resources and open-source software libraries enables faster development for new domains and languages.", "sentence2": "lT is highly language dependent and it takes considerable resources to develop lT for new languages.", "label": "contrasting"} {"id": "test_841", "sentence1": "Because of the exploratory nature of the project and the type of information that has been collected, the language actor documentation is very detailed and often shaped by the organization and work environment of the specific institution.", "sentence2": "some good candidates for facets did emerge from the data when we analyzed it specifically with this aim in mind.", "label": "contrasting"} {"id": "test_842", "sentence1": "A straightforward approach would be to share the character level vocabulary between CJK languages, as it was possible between Chinese and Japanese.", "sentence2": "this, unfortunately, is not a straightforward operation, as Hangul (the Korean writing system) is phonetic, unlike the other two examples.", "label": "contrasting"} {"id": "test_843", "sentence1": "The elaborate deep learning models created new standards in OCR.", "sentence2": "like any machine learning method, deep learning models also need training material.", "label": "contrasting"} {"id": "test_844", "sentence1": "Like kraken, it allows the user to specify the structure of the neural network with VGSL.", "sentence2": "to kraken, Tesseract is not GPU-enabled.", "label": "contrasting"} {"id": "test_845", "sentence1": "It became evident during our work that character error rates (CER) are a good indicator about the models\u2019 ability to identifying characters correctly.", "sentence2": "for any further data processing which may include indexing or applying text mining techniques, the bag-of-words F1-measure provides a better picture of the systems' performances.", "label": "contrasting"} {"id": "test_846", "sentence1": "Also, the Opus dataset is a much widely used parallel corpus resource in various researcher's works.", "sentence2": "we observed that in both of these well-known parallel resources there are many repeated sentences, which may results into the wrong results (can be higher or lower) after dividing into train, validation, and test sets, as many of the sentences, occur both in train and test sets.", "label": "contrasting"} {"id": "test_847", "sentence1": "Regardless of the MT approach applied, a MT system automatically generates an equivalent version (in some target language) of an input sentence (in some source language).", "sentence2": "despite the huge effort of the MT community, it is not possible yet to generate a perfect completely automatic translation for unrestricted domains.", "label": "contrasting"} {"id": "test_848", "sentence1": "Finding definite references and their antecedents in the coreference resolution data is easy.", "sentence2": "as we described in the experiment section, it is difficult to make the correct answer rate be 50%, because most articles can be predicted using language models.", "label": "contrasting"} {"id": "test_849", "sentence1": "For pro-drop languages like Japanese and Chinese, zero pronoun was known to be one of the most difficult problems and many specific extensions for baseline translation methods have been discussed in previous research (Taira et al., 2012;Kudo et al., 2014;Takeno et al., 2016;Wang et al., 2016;Wang et al., 2018).", "sentence2": "it seems that contextaware neural machine translation can handle Japanese zero pronouns just as effectively as overt pronouns in Englishto-Russian translation (Voita et al., 2018).", "label": "contrasting"} {"id": "test_850", "sentence1": "They built a large-scale test set from German-English bilingual texts using coreference resolution and word alignment tools.", "sentence2": "to build a large-scale test set for Japanese zero pronouns, we have to develop accurate tools for Japanese empty category detection (zero pronoun identification) and Japanese coreference resolution, which remains open problems.", "label": "contrasting"} {"id": "test_851", "sentence1": "The results of the 6th WAT suggest that most sentences that are typical in TDDC and do not depend on context are translated correctly.", "sentence2": "there are mistranslations in sentences that contain words that are not present in TDDC or whose meaning changes depending on the context.", "label": "contrasting"} {"id": "test_852", "sentence1": "The most comparable resource to the one presented here is the COPPA Corpus version 2 which contains around 13 million sentences.", "sentence2": "for other language pairs our corpus is larger, e.g. for there are 6.6M English/Japanese sentences while the JW300 corpus (Agic\u00b4 and Vulic, 2019) contains around 2.1M.", "label": "contrasting"} {"id": "test_853", "sentence1": "As shown in Section 2.1, the standard NMT usually models a text by considering isolated sentences based on a strict assumption that the sentences in a text are independent of one another.", "sentence2": "disregarding dependencies across sentences will negatively affect translation outputs of a text in terms of discourse properties.", "label": "contrasting"} {"id": "test_854", "sentence1": "MADAMIRA (Pasha et al., 2014) combines MADA (Morphological Analysis and Disambiguation of Arabic) which is built on SAMA (Standard Arabic Morphological Analyser) and AMIRA (a morphological system for colloquial Egyptian Arabic).", "sentence2": "to MADAMIRA, FSAM's rule-based system focuses on MSA templatic morphological analysis yielding root and pattern, generation and diacritization.", "label": "contrasting"} {"id": "test_855", "sentence1": "On composing both FSTs, a weighted FST mapping surface form to weighted lexical forms will be generated.", "sentence2": "if the second FST doesn't have a path for a certain analysis then the surface-form:analysis pair will be dropped.", "label": "contrasting"} {"id": "test_856", "sentence1": "For example, named entities, cognates/loanwords, and morphologically complex words that contain multiple morphemes are extremely challenging to properly tokenise because the occurrences of such terms are rare even in large training datasets.", "sentence2": "substrings of such terms are likely to be more frequent.", "label": "contrasting"} {"id": "test_857", "sentence1": "Modern Standard Arabic (MSA), the official language of the Arab world, is well studied in NLP and has an abundance of resources including corpora and tools.", "sentence2": "most Arabic dialects are considered under-resourced, with the exception of Egyptian Arabic (EGY).", "label": "contrasting"} {"id": "test_858", "sentence1": "Such models are well equipped to model some aspects of morphology implicitly as part of an end-to-end system without requiring explicit feature engineering.", "sentence2": "these models are very data-intensive, and do not scale down well in the case of low-resource languages.", "label": "contrasting"} {"id": "test_859", "sentence1": "The different analyzers provide minor or no improvements over the Neural Joint Model alone when embedding the candidate tags.", "sentence2": "the ranking approach reduces the accuracy drastically for different combinations of analyzers.", "label": "contrasting"} {"id": "test_860", "sentence1": "Indeed, in some languages (e.g., Basque) 100 words cannot cover even a single verb paradigm.", "sentence2": "even in such restricted conditions some systems perform significantly better than others, the state-of-the-art approach is imitation learning via minimization of Levenshtein distance between the network output and the correct word form (Makarov and Clematide, 2018b).", "label": "contrasting"} {"id": "test_861", "sentence1": "They were also ex\u0002tensively used in Najafi et al. (2018) system, that took the second place in Sigmorphon 2018 Shared Task.", "sentence2": "they utilized the complete Unimorph data, which is sufficiently more than 1000 word forms used in our work.", "label": "contrasting"} {"id": "test_862", "sentence1": "By restricting transformations to orthogonal linear mappings, VecMap and MUSE rely on the assumption that the monolingual embeddings spaces are approximately isomorphic (Barone, 2016).", "sentence2": "it has been argued that this assumption is overly restrictive, as the isomorphism assumption is not always satisfied (S\u00f8gaard et al., 2018;.", "label": "contrasting"} {"id": "test_863", "sentence1": "Not further improving the results for German, Czech and Italian languages might be because of the sufficiency of the target embedding usage for domain adaptation (Jurafsky and Martin, 2014) in those.", "sentence2": "the fact that Spanish and French performances improved on both Wikipedia and Twitter domains when Dom-Drift is used, might show that the necessity of DomDrift can be related to certain property of the target language, which can be further explored as future work.", "label": "contrasting"} {"id": "test_864", "sentence1": "Moreover, Turkish, Russian, Tigrigna, Polish, Uyghur, Croatian, Wolaytta, Bulgarian, German, Swedish are also characterized by high OOV rates.", "sentence2": "mandarin, Thai, Hausa, Japanese, Vietnamese and English are characterized by low OOV rate.", "label": "contrasting"} {"id": "test_865", "sentence1": "Frame-semantic parsers, however, are normally trained on manually annotated resources such as the FrameNet corpus (Baker et al., 1998) or the OntoNotes corpus (Pradhan and Xue, 2009;Weischedel et al., 2013).", "sentence2": "such annotations only exist for a small subset of the world's languages.", "label": "contrasting"} {"id": "test_866", "sentence1": "For both setups, we normalize labels to be conform with the PropBank (Palmer et al., 2005) notation (e.g., A1 becomes ARG1).", "sentence2": "as shown in Figure 1, the experiments with the full label set have a slightly better accuracy than the ones with a simplified label set, so we will present only the results for the former.", "label": "contrasting"} {"id": "test_867", "sentence1": "Given that CS is language-dependent, a corpus for each language pair is needed.", "sentence2": "collecting CS corpora is a very challenging task, thus the collected, and available, corpora are very scarce and cover few language pairs.", "label": "contrasting"} {"id": "test_868", "sentence1": "The above-mentioned differences are statistically significant.", "sentence2": "relatively large standard deviations of the metrics should be taken into account.", "label": "contrasting"} {"id": "test_869", "sentence1": "More specifically, we exploit the fact that Twitter users can make use of Twitter screen names (e.g., @UserScreen-Name) in their tweet posts to mention other users, which provides us with unambiguous mentions.", "sentence2": "e ob\u0002serve that many tweets also contain proper names (e.g. last names or acronyms for organizations) to refer to other Twitter users, thereby creating ambiguities about the user (entity) they refer to.", "label": "contrasting"} {"id": "test_870", "sentence1": "Note that we do not claim that all multimedia analysis work adopts an overly simplistic conceptualization of how text and images relate.", "sentence2": "we find the lack of research on realistic connections between text and images is serious enough that it may hold back the state of the art in multimedia analysis for disaster management.", "label": "contrasting"} {"id": "test_871", "sentence1": "In sum, our analysis reveals that the image caption to some extent describes the content of the image.", "sentence2": "in many cases the caption provides additional information which is not conveyed by the image alone.", "label": "contrasting"} {"id": "test_872", "sentence1": "Since the news articles in our collection are news reports, rather than editorials or feature articles, recency of what is reported in the text and depicted in the accompanying images is to be expected.", "sentence2": "there is also a fair proportion where the image is less recent.", "label": "contrasting"} {"id": "test_873", "sentence1": "Up until this point, we have investigated temporal distance.", "sentence2": "we have an important point to make about spatial distance that emerged from our manual analysis.", "label": "contrasting"} {"id": "test_874", "sentence1": "Success at that task could ultimately also form the basis of an automatic annotation in the future.", "sentence2": "here we limit ourselves to a pilot, which provides a basic demonstration that the categories of news articles and images can be automatically distinguished.", "label": "contrasting"} {"id": "test_875", "sentence1": "Further, articles about ongoing flooding often are associated with flood-related images.", "sentence2": "it is not advisable to assume that a flood-related image will directly relate to a flooding-event described in the corresponding article.", "label": "contrasting"} {"id": "test_876", "sentence1": "As in them, we constructed our questions and answers based on both textual and visual cues from short video clips.", "sentence2": "unlike them, our proposed dataset relies on video clips that were recorded naturally by people, without predefined scripts.", "label": "contrasting"} {"id": "test_877", "sentence1": "This insight is in line with our previous work (Schulte im Walde et al., 2016) which also demonstrated that empirical modifier properties do not have a consistent effect on the quality of predicting compound compositionality.", "sentence2": "ranges zooming into the prediction results for compounds with high-, mid and low-productivity heads (see Table 6), we do observe patterns for compound subsets.", "label": "contrasting"} {"id": "test_878", "sentence1": "This is probably one of the reasons why many studies that investigated idiomatic expressions, only collected limited information about idiom properties for very small numbers of idioms only.", "sentence2": "this is problematic for research, because it hinders comparability of results.", "label": "contrasting"} {"id": "test_879", "sentence1": "On the other hand, in concrete applications, crucial domain-specific entities need to be identified in a reliable way, such as designations of legal norms and references to other legal documents (laws, ordinances, regulations, decisions, etc.).", "sentence2": "most NER solutions operate in the general or news domain, which makes them inapplicable to the analysis of legal documents (Bourgonje et al., 2017;Rehm et al., 2017).", "label": "contrasting"} {"id": "test_880", "sentence1": "We hypothesize that an increase in training data would yield better results for BiLSTM-CRF but not outperform transfer learning approach of MTL (or even BioBERT).", "sentence2": "to other common NER corpora, like CoNLL 2003 14 , even the best baseline system only achieves relatively low scores.", "label": "contrasting"} {"id": "test_881", "sentence1": "It connects sentiment analysis and Natural Language Generation (Zhang et al., 2018a) and facilitates a lot of NLP applications such as fighting against offensive language in social media (Santos et al., 2018), news rewriting, and building controllable dialogue systems.", "sentence2": "this task is difficult in practice due to the lack of parallel data (sentences with similar content but different sentiments).", "label": "contrasting"} {"id": "test_882", "sentence1": "It suggests that the semantic representation is also essential to preserve content.", "sentence2": "the lack of semantic representation brings little decrease in sentiment transfer accuracy.", "label": "contrasting"} {"id": "test_883", "sentence1": "Natural Language Processing (NLP) can help unlock the vast troves of unstructured data in clinical text and thus improve healthcare research.", "sentence2": "a big barrier to developments in this field is data access due to patient confidentiality which prohibits the sharing of this data, resulting in small, fragmented and sequestered openly available datasets.", "label": "contrasting"} {"id": "test_884", "sentence1": "Natural Language Processing (NLP) has enormous potential to advance many aspects of healthcare by facilitating the analysis of unstructured text (Esteva et al., 2019).", "sentence2": "a key obstacle to the development of more powerful NLP methods in the clinical domain is a lack of accessible data.", "label": "contrasting"} {"id": "test_885", "sentence1": "We cannot say for certain whether using GPT-2 or EDA could positively impact our results.", "sentence2": "it appears that our EDA baseline generally performs even worse than our Transformer and GPT-2 augmentations and the Original data itself for both MimicText-98 and MimicText-9.", "label": "contrasting"} {"id": "test_886", "sentence1": "It is our hypothesis that these inaccuracies can provide an optimal amount of noise when using a model that has been pretrained on biomedical texts, thus allowing them to better generalise.", "sentence2": "this noise proves too much for models that have only been pretrained on non-medical text.", "label": "contrasting"} {"id": "test_887", "sentence1": "This leads us to hypothesise that this task might be too easy and that even weaker models are able to relatively accurately identify the phenotypes of patients from their discharge summaries.", "sentence2": "we still note that our baseline models report the highest values across our metrics, especially our 'Original' data using the BioBERT model which reports the best accuracy and F1 scores.", "label": "contrasting"} {"id": "test_888", "sentence1": "Recently, contextualized word embeddings such as BERT (Devlin et al., 2019) have largely improved the performance of NLP tasks compared to static embeddings such as word2vec (Mikolov et al., 2013), GloVe (Pennington et al., 2014).", "sentence2": "static embeddings are still frequently used to various studies for sentence embeddings (Yang et al., 2019;Almarwani et al., 2019) and even the other domains such as the extraction of interaction between drugs in the biomedical field (Sun et al., 2019).", "label": "contrasting"} {"id": "test_889", "sentence1": "These approaches effectively handle surface-variation of words such as inflected forms and typos.", "sentence2": "this is only feasible when the roots of words are existed in a vocabulary.", "label": "contrasting"} {"id": "test_890", "sentence1": "The authors show that CamemBERT obtains significant improvements on many French tasks compared to the publicly available multilingual BERT.", "sentence2": "the architecture used in CamemBERT is different to BERT, which makes the comparison among the models less straightforward.", "label": "contrasting"} {"id": "test_891", "sentence1": "Such results show that mean smiling intensity seems to be a more robust criterium than the sole presence or absence of smile to evaluate the impact of smiling on the success or failure of humor.", "sentence2": "based on only four participants including one exception (MA), such a result cannot be more precise.", "label": "contrasting"} {"id": "test_892", "sentence1": "This result could indicate that mean smiling intensity has a larger impact of the success or failure of humor than the simple presence of smiling.", "sentence2": "such a result cannot be considered meaningful either because, this only holds for three participants.", "label": "contrasting"} {"id": "test_893", "sentence1": "For example, an NMT system might translate the source sentence given in (1) as the viable Spanish translation given in (2).", "sentence2": "in a given organizational context (3) or even (4) might instead be the approved translation.", "label": "contrasting"} {"id": "test_894", "sentence1": "The current implementation only uses a flat list of nonhierarchical source-target term pairs (often referred to as a \"glossary\") to seed the injection process, which increases the likelihood of polysemic collisions.", "sentence2": "terminology management practices involve much more complex termbases that express several types of lexical, domain, semantic, hierarchical (taxonomic), ontological, overlapping, and nesting relationships among terms.", "label": "contrasting"} {"id": "test_895", "sentence1": "Overall, the percentage of rare words gets smaller as corpus size increases, as more and more words appear over 10 times.", "sentence2": "the hyperparameters seem to have different effects on this value depending on corpus size as well.", "label": "contrasting"} {"id": "test_896", "sentence1": "The naming form distribution in the GTTC does not reflect the natural distribution of naming forms because we oversampled Dr.-containing tweets.", "sentence2": "the relation between naming and stance should not be affected by oversampling.", "label": "contrasting"} {"id": "test_897", "sentence1": "Finally, the political orientation of users was determined by a proxy that assumes that if you criticise right-wing politicians you are more likely to be left-leaning and vice versa.", "sentence2": "of course, you can also criticise politicians from your own political spectrum.", "label": "contrasting"} {"id": "test_898", "sentence1": "A sample of titlecontaining tweets suggests that titles are not a definite signal that an otherwise positive-sounding German tweet is meant to be interpreted negatively (or vica versa).", "sentence2": "since the use of honorifics can be indicative of sarcasm (Liu et al., 2014), it is worth investigating whether the use of titles alongside explicit negative stance should be interpreted as sarcasm, and whether this sarcastic use plays a role in causing the weaker positive association with formal naming in left-leaning discourse.", "label": "contrasting"} {"id": "test_899", "sentence1": "A typical application of classical SA in an industrial setting would be to classify a document like a product review into positive, negative or neutral sentiment polarity.", "sentence2": "to SA, the more fine-grained task of Aspect-Based Sentiment Analysis (ABSA) (Hu and Liu, 2004;Pontiki et al., 2015) aims to find both the aspect of an entity like a restaurant, and the sentiment associated with this aspect.", "label": "contrasting"} {"id": "test_900", "sentence1": "As we mentioned above, majority works have boiled down opinion mining to a problem of classifying whether a piece of text expresses positive or negative evaluation.", "sentence2": "evaluation is, in fact, much more complex and multifaceted, which varies depending on linguistic factors, as well as participants of the communicative activity.", "label": "contrasting"} {"id": "test_901", "sentence1": "Many polarity shifters can affect both positive and negative polar expressions, shifting them towards the opposing polarity.", "sentence2": "other shifters are restricted to a single shifting direction.", "label": "contrasting"} {"id": "test_902", "sentence1": "The dataset created in (Castro et al., 2018) was used by Castro et al. (2018) in the context of the HAHA 2018 competition for Humor Detection and Funniness Average Prediction", "sentence2": "the dataset presented in (Castro et al., 2018) still presents some issues.", "label": "contrasting"} {"id": "test_903", "sentence1": "None of the systems could beat this baseline in terms of precision.", "sentence2": "the recall of this baseline is very low, because many humorous tweets are not written as dialogues, and that is why its F1 score is not that high.", "label": "contrasting"} {"id": "test_904", "sentence1": "For a further development of this line of work, it is essential to construct a linguistically valid treebank on CG.", "sentence2": "current corpora based on CG often do not take advantage of linguistically adequate analyses developed in the CG literature, mainly because these corpora are converted from existing resources which do not contain fine-grained annotation (Honnibal et al., 2010).", "label": "contrasting"} {"id": "test_905", "sentence1": "Information about a predicate's arguments are encoded only indirectly and their immediate accessibility depends on the precise type of PS treebank.", "sentence2": "dependency structures (DS) abstract away from linear order and concentrate on encoding functional dependencies between the items of a clause.", "label": "contrasting"} {"id": "test_906", "sentence1": "One can include functional labels representing grammatical relations in a PS parser.", "sentence2": "it has been shown that training a PS parser by including functional labels produces lower constituency parsing accuracy.", "label": "contrasting"} {"id": "test_907", "sentence1": "It includes an Urdu text corpus of 1.6 million words and a parallel English-Urdu corpus containing 200K words.", "sentence2": "the Urdu EMILLE copora are unannotated with respect to grammatical structure.", "label": "contrasting"} {"id": "test_908", "sentence1": "It seems that the main takeaway is \"the more, the better,\" as the top-scoring setup uses all five auxiliary treebanks.", "sentence2": "we get a significantly stronger improvement from the constituency treebanks than from the dependency treebanks.", "label": "contrasting"} {"id": "test_909", "sentence1": "For language-specific models and questions, such representations are often adequate and may even be preferable to the alternatives.", "sentence2": "in multilingual models, the language-specific nature of phonemic abstractions can be a liability.", "label": "contrasting"} {"id": "test_910", "sentence1": "The usefulness of AlloVera for all purposes will increase as it grows to cover a broad range of the languages for which phonetic and phonological descriptions have been completed.", "sentence2": "to illustrate the usefulness of AlloVera, we will rely primarily on the zero-shot, universal ASR use-case in the evaluation in this paper.", "label": "contrasting"} {"id": "test_911", "sentence1": "Then phonetic representations could be derived by first applying G2P to the orthographic text, then applying the appropriate transducer to the resulting phonemic representation.", "sentence2": "constructing such a resource is expensive, requires several specialized skills on the part of curatorswho must encode the phonological environments in which allophones occur-and requires information that is often omitted from phonological descriptions of languages.", "label": "contrasting"} {"id": "test_912", "sentence1": "Therefore, adding more languages can confuse the model, leading it to assign incorrect phonemes.", "sentence2": "alloVera provides a consistent assignment across languages by using allophone inventories.", "label": "contrasting"} {"id": "test_913", "sentence1": "The universal phone inventory consists of all allophones in AlloVera.", "sentence2": "the shared phoneme model could only generate inconsistent universal phonemes and the private phoneme model could only generate languagespecific phonemes.", "label": "contrasting"} {"id": "test_914", "sentence1": "The term allophone was coined by Benjamin Lee Whorf in the 1920s and was popularized by Trager and Block (1941).", "sentence2": "the idea goes back much further, to Baudoin de Courtenay (1894).", "label": "contrasting"} {"id": "test_915", "sentence1": "The row tagged with Full means that the whole training set was used to train the multilingual model.", "sentence2": "the row with tag Low is trained under a low resource condition in which we only select 10k utterances from each training corpus.", "label": "contrasting"} {"id": "test_916", "sentence1": "The recognition and automatic annotation of temporal expressions (e.g. Add an event for tomorrow evening at eight to my calendar) is a key module for AI voice assistants, in order to allow them to interact with apps (for example, a calendar app).", "sentence2": "in the NLP literature, research on temporal expressions has focused mostly on data from the news, from the clinical domain, and from social media.", "label": "contrasting"} {"id": "test_917", "sentence1": "Snips (Coucke et al., 2018) 5 is a crowdsourced dataset for the voice assistant domain, specifically for seven intents 6 , which is widely used for benchmarking NLU components of voice assistants.", "sentence2": "no explicit details are provided on how the data was created or collected, and it does not appear to come from a real-world interaction with a voice assistant (sentences from Snips can at times be rather odd, albeit grammatical, e.g.", "label": "contrasting"} {"id": "test_918", "sentence1": "An hour is then annotated as a DURATION.", "sentence2": "the reservation needs to be done for a specific punctual time that is not expressed here.", "label": "contrasting"} {"id": "test_919", "sentence1": "Similar to Wang et al. 2019, on the one hand we annotate the reliance on factual knowledge, that is (Geo)political/Legal, Cultural/Historic, Technical/Scientific and Other Domain Specific knowledge about the world that can be expressed as a set of facts.", "sentence2": "we denote Intuitive knowledge requirements, which is challenging to express as a set of facts, such as the knowledge that a parenthetic numerical expression next to a person's name in a biography usually denotes his life span.", "label": "contrasting"} {"id": "test_920", "sentence1": "For the non-BERT models, DocQA utilizes the loss value from the answer span prediction to check answerability, while Read+Verifier introduces a new classifier for verifying the question and answer pair.", "sentence2": "bERTbased models first pretrain deep bidirectional representations from large-scale unlabeled text without any explicit modeling for a specific task.", "label": "contrasting"} {"id": "test_921", "sentence1": "These previous studies have mostly focused on factoid questions, each of which can be answered in a few words or phrases generated by understanding multimodal contents in a short video clip.", "sentence2": "this problem definition of video question answering causes some practical limitations for the following reasons.", "label": "contrasting"} {"id": "test_922", "sentence1": "Beyond these functionalities, which can in principle be achieved with large finite-state dialogue models, commercial assistants were not intended to support extended conversations on topics in the general domain, although their capacities for chitchat have been gradually increasing (Fang et al., 2018).", "sentence2": "the systems known as chatbots were intended from the start to offer robust conversational capacities, although with a thin and often implicit knowledge base.", "label": "contrasting"} {"id": "test_923", "sentence1": "Given the outstanding recent results of the BERT model applied to QA on the SQuAD dataset, we selected this model for our QA component.", "sentence2": "the model is designed for answer extraction from a specific paragraph -with provision for the cases when the paragraph does not contain the answer -while we need to develop an end-to-end solution, starting directly from the document repository, without prior knowledge of the paragraphs relevant to each question.", "label": "contrasting"} {"id": "test_924", "sentence1": "These analyses confirm the possibility to use syntactic n-gram features in cross-lingual experiments to categorize texts according to their CEFR level (Common European Framework of Reference for Languages).", "sentence2": "text length and some classical indexes of readability are much more effective in the monolingual and the multilingual experiments than what Vajjala and Rama concluded and are even the best performing features when the cross-lingual task is seen as a regression problem.", "label": "contrasting"} {"id": "test_925", "sentence1": "Most of the works in this area have been focused on English as a second language where the needs are the most obvious (Condon, 2013;Weigle, 2013).", "sentence2": "although knowing a lingua franca is important, being able to integrate oneself into another culture by communicating in its language is also important.", "label": "contrasting"} {"id": "test_926", "sentence1": "In this experiment, the texts written in German are used to predict separately the CEFR level of the Italian and Czech texts.", "sentence2": "as shown in Table 1, there are very large disparities between the three languages regarding the number of texts classified in each CEFR level.", "label": "contrasting"} {"id": "test_927", "sentence1": "The analyses support their conclusion that it is possible to use features like POStag or dependency ngrams to learn a predictive model on one language and then use it to categorize texts written in another language.", "sentence2": "the complementary analyses suggest that the number of words in each text and some classical indexes of readability are more effective when the task is seen as a regression problem.", "label": "contrasting"} {"id": "test_928", "sentence1": "It follows that making the code available in the form of a Docker image as it was required for REPROLANG 2020 is useful in order to ensure that the results can be reproduced, but it does not guarantee that these results correspond to the explanations given in the paper.", "sentence2": "providing a Docker image makes it possible to find the versions of the programs and modules used, often necessary for reproducing exactly a study.", "label": "contrasting"} {"id": "test_929", "sentence1": "These patterns are helpful to find irony samples in a given corpus.", "sentence2": "they cannot be used as irony detection algorithms due to their limited coverage.", "label": "contrasting"} {"id": "test_930", "sentence1": "We initially attempted to set this up as a labeling task by providing participants labels and definitions.", "sentence2": "we found it difficult for annotators to pinpoint differences between versions where edits modify or provide new information, in contrast to providing only stylistic changes.", "label": "contrasting"} {"id": "test_931", "sentence1": "They are considered improvements over word models, and their effectiveness is usually judged with benchmarks such as semantic similarity datasets.", "sentence2": "most of these datasets are not designed for evaluating sense embeddings.", "label": "contrasting"} {"id": "test_932", "sentence1": "As mentioned in Section 3.2, GenSense-1 and SenseRetro-1, in which only first sense vectors are utilized, outperform their multi-sense counterparts in MEN, RW, WS353, and SCWS.", "sentence2": "on MSD-1030 an opposite pattern is shown.", "label": "contrasting"} {"id": "test_933", "sentence1": "Thus, semantic attribute vectors can effectively capture the commonalities and differences among concepts.", "sentence2": "as semantic attributes have been generally created by psychological experimental settings involving human annotators, an automatic method to create or extend such resources is highly demanded in terms of language resource development and maintenance.", "label": "contrasting"} {"id": "test_934", "sentence1": "Interestingly, the predicted attributes (originated from visual attributes annotated in ViSA) contributed to the performance improvement in VisSim and SemSim, outperforming not only the existing results but also that with fast-Text.", "sentence2": "significant differences were not observed between the dir and ext tasks.", "label": "contrasting"} {"id": "test_935", "sentence1": "So far, brain research has offered insight on the fact that we develop mental images for words learned.", "sentence2": "what this fails to tell us is how they would look like visually.", "label": "contrasting"} {"id": "test_936", "sentence1": "Abstract shapes contain descriptions like \"rectangle\" for Turkey or \"pointy\" for Somalia.", "sentence2": "associations use other objects that look similar to the shape of the country as a reference.", "label": "contrasting"} {"id": "test_937", "sentence1": "This shows that people's size description generally matches the actual size of the country on the world map.", "sentence2": "the location of the country seems to play an important role in the accuracy of people's descriptions.", "label": "contrasting"} {"id": "test_938", "sentence1": "These datasets allow modern machine learning techniques to glean insight from the massive amounts of textual data they contain.", "sentence2": "in the areas of humor classification and generation we find much smaller datasets, due to the complexity of humorous natural language.", "label": "contrasting"} {"id": "test_939", "sentence1": "A task similar to that of (Hossain et al., 2019;Weller and Seppi, 2019) can be done with this dataset, where a model predicts the level of humor found in the joke in order to examine what characterizes humor.", "sentence2": "due to Reddit's large scale and uneven distribution of upvotes, predicting the number of upvotes would be a sparse and difficult task.", "label": "contrasting"} {"id": "test_940", "sentence1": "Many studies have been proposed in recent years to deal with online abuse, where swear words have an important role, providing a signal to spot abusive content.", "sentence2": "as we can expect observing the different facets of swearing in social environments, the presence of swear words could also lead to false positives when they occur in a nonabusive context.", "label": "contrasting"} {"id": "test_941", "sentence1": "Zannettou et al. (2018) compared the behavior of troll accounts with a random set of Twitter users to analyze the influence of troll accounts on social media.", "sentence2": "new accounts can be opened at any time, and troll accounts can be suspended or deleted at any time.", "label": "contrasting"} {"id": "test_942", "sentence1": "As such, we can conclude that these selected stylometric features can be successfully transferred from one language to another.", "sentence2": "most of the stylometric features are language-dependent and will also rely on external natural language processing techniques.", "label": "contrasting"} {"id": "test_943", "sentence1": "In tweet (7), the writer uses a joke to make an ironic statement about the social problem of the reluctance of young men to marry.", "sentence2": "in example (8) and despite the use of the hashtag (\"#irony\"), the tweet is just a simple joke about a girl mosquito.", "label": "contrasting"} {"id": "test_944", "sentence1": "The only previous attempt at normalizing Italian social media data is from Weber and Zhekova (2016).", "sentence2": "they have a different scope of the task, mostly focusing on readability, not on normalization on the lexical level.", "label": "contrasting"} {"id": "test_945", "sentence1": "However, perhaps surprisingly, when training on canonical data (ISDT), using predicted normalization on the input data leads to a slightly better performance compared to using gold.", "sentence2": "the differences are very minor in this setting, and considering the size of the test data (100 tweets), we can not draw any conclusions from these results.", "label": "contrasting"} {"id": "test_946", "sentence1": "Prior studies have analyzed how location affects the type of language that people use, often looking at text written by authors from different countries when exploring crosscultural differences (Poblete et al., 2011; Garcia-Gavilanes et al., 2013).", "sentence2": "it is not always necessary to look at multiple countries in order to view different cultures.", "label": "contrasting"} {"id": "test_947", "sentence1": "In most cases, the speech is clearly pronounced, well articulated and easy to understand.", "sentence2": "oral history interviews are often recorded using conventional recording equipment that was common at the time of recording.", "label": "contrasting"} {"id": "test_948", "sentence1": "This setup also achieves slightly better results than the proposed approach on the Challenging Broadcast test set with the larger language model.", "sentence2": "the gain is less than 0.2% relative.", "label": "contrasting"} {"id": "test_949", "sentence1": "After launching the procedure of annotation, Analor creates another TextGrid file for every sound file containing a new tier of automatically segmented periods.", "sentence2": "analor creates only one tier with periods but TextGrid files of manual annotation contain a tier for each speaker in every sound file.", "label": "contrasting"} {"id": "test_950", "sentence1": "Works like Tacotron (Wang et al., 2017),Tacotron 2 (Shen et al., 2017), Deep Voice 3 (Ping et al., 2017) are capable of producing high quality natural speech.", "sentence2": "all of these methods are data hungry and require approximately 24 hrs of text-to-speech data for a single speaker.", "label": "contrasting"} {"id": "test_951", "sentence1": "This corpus was used to train text-to-speech systems for 13 languages were developed in (Pradhan et al., 2015).", "sentence2": "in this corpus, the amount of data provided per language is far too less (\u224825% of recent TTS datasets) for training recent neural network based systems that can produce natural, accurate speech.", "label": "contrasting"} {"id": "test_952", "sentence1": "It can also be seen that the alignment curve shape is inferior in the case of Malayalam, and this is also reflected in the lower MOS scores for Malayalam.", "sentence2": "hindi and Bengali has near perfect alignment curves which corresponds to the higher MOS scores that we get for these languages.", "label": "contrasting"} {"id": "test_953", "sentence1": "Using a limited database of Mizo tones, the authors reported that pitch height and F0 slope can automatically classify Mizo tones into the four phonological categories with considerable accuracy of 70.28%.", "sentence2": "this work had several shortcomings, firstly, the Mizo database used for the work was considerably small; secondly, the approach for identifying tones was threshold based and no statistical method was incorporated.", "label": "contrasting"} {"id": "test_954", "sentence1": "Each tone combination consists of five unique phrases which were recorded three times by each speaker which outcome is 17, 280 phrases resulted in 54, 720 total tokens (19 speakers x 64 tonal combinations x 5 trisyllabic phrases x 3 monosyllables x 3 repetitions).", "sentence2": "22, 770 tokens are not considered as these are the low tones derived from RTS which is not considered in the present work.", "label": "contrasting"} {"id": "test_955", "sentence1": "The design of our corpus is based on S-JNAS, so we also used the ATR 503 sentences and JNAS newspaper article sentences as the script for our participants.", "sentence2": "unlike S-JNAS, we used the ATR 503 sentences as training data and the newspaper article sentences as test data.", "label": "contrasting"} {"id": "test_956", "sentence1": "For the S-JNAS corpus, each of the training data speakers read aloud two sets of ATR 503 sentences (about 100 sentences) and one set of the newspaper article sentences (about 100 sentences).", "sentence2": "for our corpus, since many of our speakers are very elderly, and some have limited vision or a tendency towards dementia, we limited the number of sentenced we asked each participant to read in order to reduce the burden.", "label": "contrasting"} {"id": "test_957", "sentence1": "Unlike in the other areas, there was insufficient coaching of the Tokushima participants by the recording staff, such as prompting them to read the text more carefully when they made mistakes, or having them re-read the text aloud when they made serious errors.", "sentence2": "speech from the Yamagata speakers obtained the best recognition results, and this may have been because the average age of the participants was the lowest, at 73.4 years, which may have helped them to read aloud more fluently.", "label": "contrasting"} {"id": "test_958", "sentence1": "The result shows that when dealing with American read material, the word error rate (WER) was 3.1% and when dealing with American/Canadian spontaneous speech, WER was 7.94%.", "sentence2": "when the system was used to transcribe IE, WER was much higher and was 22.89%.", "label": "contrasting"} {"id": "test_959", "sentence1": "In general, the scores on the manual pyramids are higher than on automatic pyramids as the average scores in table 5 show.", "sentence2": "the high Pearson's correlation between quality scores on manual and automatic pyramids, especially when we use emb 2m, leads us to argue that this could be an issue of coverage.", "label": "contrasting"} {"id": "test_960", "sentence1": "Research on fact checking has benefited from large-scale datasets such as FEVER and SNLI.", "sentence2": "such datasets suffer from limited applicability due to the synthetic nature of claims and/or evidence written by annotators that differ from real claims and evidence on the internet.", "label": "contrasting"} {"id": "test_961", "sentence1": "In more extreme cases, the claim starts with a pronoun.", "sentence2": "it may be necessary to know that before determining whether the claim is supported or not, because there may be multiple lines in the original evidence source that could be talked about.", "label": "contrasting"} {"id": "test_962", "sentence1": "One way to generate refuted claims is to perform automatic claim negation using rulebased 'not' insertion based on syntax and part-of-speech (Bilu et al., 2015).", "sentence2": "this would result in a handful of negation words appearing in the refuted claims by design, causing classifiers to exploit this pattern.", "label": "contrasting"} {"id": "test_963", "sentence1": "Another automated approach is to pick a different random claim from the dataset.", "sentence2": "a \"refuted\" claim chosen this way is likely to be topically dissimilar from the evidence file, rendering it not a useful negative example for the classifier.", "label": "contrasting"} {"id": "test_964", "sentence1": "The crucial differentiating factor, then, might be the existence of antonyms: a claim is more likely to be refuted by e c if it contains even one antonym of a word in e c .", "sentence2": "a claim c that is supported by e c should tend to have no antonyms.", "label": "contrasting"} {"id": "test_965", "sentence1": "Some such data sets which have enabled the advancement of NLI (and fact verification) are SNLI (Bowman et al., 2015) MNLI (Williams et al., 2017), FEVER (Thorne et al., 2018), and FNC (Pomerleau and Rao, 2017).", "sentence2": "these datasets are not devoid of biases (subtle statistical patterns in a dataset, which could have been introduced either due to the methodology of data collection or due to an inherent social bias).", "label": "contrasting"} {"id": "test_966", "sentence1": "For example, in the second data point the bias of Clinton towards the label Agree (i.e., the percentage of data points where the entity Clinton cooccurred with the label Agree) is 63.15%.", "sentence2": "the model trained on delexicalized data was able to predict the label with a lower bias (Disagree with 36.85%).", "label": "contrasting"} {"id": "test_967", "sentence1": "In Thai, spaces are used to separate sentences.", "sentence2": "they are used for other purposes as well, such as separating phrases, clauses, and listed items.", "label": "contrasting"} {"id": "test_968", "sentence1": "To address this problem, ideally, the corpus needs to be extended to cover the target domain.", "sentence2": "this usually comes with high costs and requires time, so it is often not feasible.", "label": "contrasting"} {"id": "test_969", "sentence1": "We suppose that the language model used in this study, which is trained mainly on hotel reviews, could be the most beneficial for segmenting user-generated data in the hotel domain.", "sentence2": "there is no annotated corpus for the hotel domain publicly available at the current time.", "label": "contrasting"} {"id": "test_970", "sentence1": "Since for uttering responses with the high concreteness like \"response 6\" it is necessary to deeply consider the content of narratives, the degree of empathy shown by these responses tends to be high.", "sentence2": "since it is not necessary to deeply consider the content of narratives for uttering responses with low concreteness like \"response 5,\" the degree of empathy shown by these responses tends to be low.", "label": "contrasting"} {"id": "test_971", "sentence1": "Since responses of high versatility are uttered at various points in narrative speech, it is considered that these responses occur along with many other types of responses.", "sentence2": "since responses of low versatility are uttered in fewer points in narrative speech, it is considered that these responses do not occur along with many other types of responses.", "label": "contrasting"} {"id": "test_972", "sentence1": "Empathy to narratives encourages a speaker to speak more only when the degree of the empathy is appropriate.", "sentence2": "when the degree of the empathy is not appropriate for the narrative, such empathy discourages the speaker.", "label": "contrasting"} {"id": "test_973", "sentence1": "It offers flexibility to choose the type of annotation and labels as well as several other options during the annotation (e.g., sentence marking, break line and white space deletion).", "sentence2": "it does not support multiple parallel annotators nor it is deployed in a server, thus, lacking the ability to track the annotator's progress and to flexibly work on different machines.", "label": "contrasting"} {"id": "test_974", "sentence1": "Doccano also allows for the setting of task-specific labels.", "sentence2": "only categorical labels are supported and the customization of these is also limited to annotation tasks with similar label requirements such as NER, sentiment analysis, and translation.", "label": "contrasting"} {"id": "test_975", "sentence1": "AWOCATo includes a customizable guideline page in HTML.", "sentence2": "to cater to users with limited HTML knowledge, the annotation guidelines can be created, for example, in Google Docs 13 , exported as HTML and stored in a predefined folder.", "label": "contrasting"} {"id": "test_976", "sentence1": "This complexity in the code is the artifact of supporting a very large number of features that are needed in certain cases.", "sentence2": "there are cases where all these features might not be necessary, for instance, researcher who are new to sequence-to-sequence (seq2seq) modeling might need more simpler codes to start with.", "label": "contrasting"} {"id": "test_977", "sentence1": "ESTNLTK library is an extendable collection of NLP utilities which use Text objects to communicate with each other.", "sentence2": "practice showed that the original structure of Text objects was not easily extendable and we had to rethink how the information is stored and structured.", "label": "contrasting"} {"id": "test_978", "sentence1": "Evaluation data was initially taken from the Estonian National Corpus (ENC) (Kallas and Koppel, 2018), which is the largest published collection of Estonian texts so far.", "sentence2": "we discovered errors in one of its subcorpora.", "label": "contrasting"} {"id": "test_979", "sentence1": "More generic specifications for component metadata are currently developed in coordination with Teanga development, and will be partially based on the Fintan ontology.", "sentence2": "we plan to align our specifications also with those of the European Language Grid (ELG), 27 and thus anticipate a longer consolidation process and several cycles of revision until we arrive at stable specifications.", "label": "contrasting"} {"id": "test_980", "sentence1": "Ideally these data should record interactions between real users and a dialogue system, or, if a dialogue system is not available (which is very common in the initial stages of development), interactions between real users and a Wizard (a human playing the role of the system), in a so called Wizard of Oz (WOz) setting (Dahlb\u00c3\u0192\u00c2\u00a4ck et al., 1993).", "sentence2": "this approach can be quite expensive and time consuming.", "label": "contrasting"} {"id": "test_981", "sentence1": "Also, to take into account the fact that SU actions are generated based on a probability distribution, expected precision, expected recall, and expected accuracy are used (Georgila et al., 2006).", "sentence2": "these metrics can be problematic because if a SU action is not the same as the user action in the reference corpus, this does not necessarily mean that it is a poor action.", "label": "contrasting"} {"id": "test_982", "sentence1": "Detecting and interpreting the temporal patterns of gaze behaviour cues is natural for humans and also mostly an unconscious process.", "sentence2": "these cues are difficult for conversational agents such as robots or avatars to process or generate.", "label": "contrasting"} {"id": "test_983", "sentence1": "The Wikisource contains the full text of the Twenty-Four Histories under the Creative Commons license, i.e., they could be freely used, re-distributed, and modified.", "sentence2": "there are two limitations: the philological provenance and the current format.", "label": "contrasting"} {"id": "test_984", "sentence1": "In this paper, we look to characterize phonotactics at the language level.", "sentence2": "we use methods more typically applied to specific sentences in a language, for example in the service of psycholinguistic experiments.", "label": "contrasting"} {"id": "test_985", "sentence1": "We conclude that /x/ is in the consonant inventory of at least some native English speakers.", "sentence2": "counting it on equal status with the far more common /k/ when determining complexity seems incorrect.", "label": "contrasting"} {"id": "test_986", "sentence1": "Like final obstruent devoicing, vowel harmony plays a role in reducing the number of licit syllables.", "sentence2": "to final obstruent devoicing, however, vowel harmony acts cross-syllabically.", "label": "contrasting"} {"id": "test_987", "sentence1": "QDMR abstracts away the context needed to answer the question, allowing in principle to query multiple sources for the same question.", "sentence2": "to semantic parsing, QDMR operations are expressed through natural language, facilitating annotation at scale by non-experts.", "label": "contrasting"} {"id": "test_988", "sentence1": "QDMR is primarily inspired by SQL (Codd, 1970;Chamberlin and Boyce, 1974).", "sentence2": "while SQL was designed for relational databases, QDMR also aims to capture the meaning of questions over unstructured sources such as text and images.", "label": "contrasting"} {"id": "test_989", "sentence1": "We see compression in both types of contexts, which suggests that the cognitive load hypothesis is the more likely account.", "sentence2": "these two hypotheses are not mutually exclusive.", "label": "contrasting"} {"id": "test_990", "sentence1": "In these methods, syntactic-guidance is sourced from a separate exemplar sentence.", "sentence2": "this prior work has only utilized limited syntactic information available in the parse tree of the exemplar sentence.", "label": "contrasting"} {"id": "test_991", "sentence1": "Our task is similar in spirit to Iyyer et al. (2018) and Chen et al. (2019a), which also deals with the task of syntactic paraphrase generation.", "sentence2": "the approach taken by them is different from ours in at least two aspects.", "label": "contrasting"} {"id": "test_992", "sentence1": "Recent studies (e.g., Linzen et al., 2016; Marvin and Linzen, 2018; have explored this question by evaluating LMs' preferences between minimal pairs of sentences differing in grammatical acceptability, as in Example 1.", "sentence2": "each of these studies uses a different set of metrics, and focuses on a small set of linguistic paradigms, severely limiting any possible bigpicture conclusions.", "label": "contrasting"} {"id": "test_993", "sentence1": "Marvin and Linzen (2018) expand the investigation to negative polarity item and reflexive licensing.", "sentence2": "these and related studies cover a limited set of phenomena, to the exclusion of well-studied phenomena in linguistics such as control and raising, ellipsis, quantification, and countless others.", "label": "contrasting"} {"id": "test_994", "sentence1": "in the simple LM method, suggesting that the probabilities Transformer-XL assigns to the irrelevant part at the end of the sentence very often overturn the observed preference based on probability up to the critical word.", "sentence2": "gPT-2 benefits from reading the whole sentence for BINDINg phenomena, as its performance is better in the simple LM method than in the prefix method.", "label": "contrasting"} {"id": "test_995", "sentence1": "Our benchmarks are exact reproducible in the sense that we provide the tables that record all model results (Section 3.3) and the code to run and evaluate our HPO algorithms (Section 6).", "sentence2": "they are not guaranteed to be broad reproducible, because the generalizability of the results might be restricted due to fixed collections of hyperparameter configurations, the variance associated with multiple runs, and the unknown best representative set of MT data.", "label": "contrasting"} {"id": "test_996", "sentence1": "The preceding discussion shows that if we have a set of terminals that are anchors for the true nonterminals in the original grammar, then the productions and the (bottom-up) parameters of the associated productions will be fixed correctly, but it says nothing about parameters that might be associated to productions that use other nonterminals.", "sentence2": "it is easy to show that under these assumptions there can be no other nonterminals.", "label": "contrasting"} {"id": "test_997", "sentence1": "One strand of research looks at using the IO algorithm to train some heuristically initialized grammar (Baker, 1979;Lari and Young, 1990;Pereira and Schabes, 1992;de Marcken, 1999).", "sentence2": "this approach is only guaranteed to converge to a local maximum of the likelihood, and does not work well in practice.", "label": "contrasting"} {"id": "test_998", "sentence1": "For example, if Bob is a bakeoff organizer, he might want accuracy above 60% in order to determine whether to manually check the submission.", "sentence2": "if Bob is providing ''MT as a service'' with strong privacy guarantees, he may need to provide the client with accuracy higher than 90%.", "label": "contrasting"} {"id": "test_999", "sentence1": "Equipped with external knowledge and multi-task learning, our model can further reduce chaotic logic and meanwhile avoid repetition.", "sentence2": "the analysis result illustrates that generating a coherent and reasonable story is challenging.", "label": "contrasting"} {"id": "test_1000", "sentence1": "But the model does not succeed in the style transfer task, and simply learns to add the word doctors into layman sentences while almost keeping the other words unchanged; and adding the word eg into the expertise sentences.", "sentence2": "it achieves good performance on all of the three ST measures, but makes little useful modifications.", "label": "reasoning"} {"id": "test_1001", "sentence1": "Moreover, these two structure encoders are bidirectionally calculated, allowing them to capture label correlation information in both top-down and bottom-up manners.", "sentence2": "hiAGM is more robust than previous top-down models and is able to alleviate the problems caused by exposure bias and imbalanced data.", "label": "reasoning"} {"id": "test_1002", "sentence1": "Note that DAG can be converted into a tree-like structure by distinguishing each label node as a single-path node.", "sentence2": "the taxonomic hierarchy can be simplified as a tree-like structure.", "label": "reasoning"} {"id": "test_1003", "sentence1": "The key information pertaining to text classification could be extracted from the beginning statements.", "sentence2": "we set the maximum length of token inputs as 256.", "label": "reasoning"} {"id": "test_1004", "sentence1": "Regardless of their source, emails are usually unstructured and difficult to process even for human readers (Sobotta, 2016).", "sentence2": "many approaches have been proposed for cleansing newsgroup and email data.", "label": "reasoning"} {"id": "test_1005", "sentence1": "Importantly, dependency paths between content words do not generally contain function words.", "sentence2": "by comparing paths across languages, differences in the surface realization are often masked, and argument structure and linkage differences emphasized.", "label": "reasoning"} {"id": "test_1006", "sentence1": "Japanese and Korean are largely similar from the point of view of language typology (SOV word order, topic prominence, agglutinative morphol\u0002ogy), but there are also important differences on the level of usage.", "sentence2": "the adjective class in Korean is less productive, and translations often resort to relative clauses for the purposes of nominal modification.", "label": "reasoning"} {"id": "test_1007", "sentence1": "The \"most common other path\" for both Russian and French is xcomp+nsubj, which is easy to explain: PUD corpora of these languages \"demote\" fewer auxiliary predicates than English (criteria for demotion are formulated in terms of superficial syntax and differ between languages) and more often place the dependent predicates as the root.", "sentence2": "in constructions like he could do something the direct edge between the subject and the verb of the dependent clause is replaced with two edges going through the modal predicate.", "label": "reasoning"} {"id": "test_1008", "sentence1": "This solves the issue of complex categorical modeling but makes slot-filling dependent on an intent detector.", "sentence2": "we propose a framework that treats slot-filling as a fully intentagnostic span extraction problem.", "label": "reasoning"} {"id": "test_1009", "sentence1": "Second, the train-test splits of NICHE dataset contain same CNs since the splitting has been done using one paraphrase for each HS and its all original CNs, while CROWD train-test splits have a similar property since an exact same CN can be found for many different HSs.", "sentence2": "the non-pretrained transformer models, which are more prone to generating an exact sequence of text from the training set, show a relatively better performance with the standard metrics in comparison to the advanced pre-trained models.", "label": "reasoning"} {"id": "test_1010", "sentence1": "In fact, we observed that, after the output CN, the over-generated chunk of text consists of semantically coherent brand-new HS-CN pairs, marked with proper HS/CN start and end tokens consistent with the training data representation.", "sentence2": "on top of CN generation for a given HS, we can also take advantage of the over-generation capabilities of GPT-2, so that the author module can continuously output plausible HS-CN pairs without the need to provide the HS to generate the CN response.", "label": "reasoning"} {"id": "test_1011", "sentence1": "Although it seems more intuitive to focus on precision since we search for an effective filtering over many possible solutions, we observed that a model with a very high precision tends to overfit on generic responses, such as \"Evidence please?\".", "sentence2": "we aim to keep the balance between the precision and recall and we opted for F1 score for model selection.", "label": "reasoning"} {"id": "test_1012", "sentence1": "The main goal of our effort is to reduce the time needed by experts to produce training data for automatic CN generation.", "sentence2": "the primary evaluation measure is the average time needed to obtain a proper pair.", "label": "reasoning"} {"id": "test_1013", "sentence1": "We performed some manual analysis of the selected CNs and we observed that especially for the Reviewer\u22652 case (which was the most problematic in terms of RR and novelty) there was a significantly higher ratio of \u201cgeneric\u201d responses, such as \u201cThis is not true.\u201d or \u201cHow can you say this about an entire faith?\u201d, for which reviewers agreement is easier to attain.", "sentence2": "the higher agreement on the generic CNs reveals itself as a negative impact in the diversity and novelty metrics.", "label": "reasoning"} {"id": "test_1014", "sentence1": "In this case, generalizability of evaluation results becomes questionable.", "sentence2": "our evaluation methodology needs to fulfill the following two requirements: (1) evaluation must not be performed on translational equivalents of the Source entries to which the model already had access during training (e.g., Sonnenschein and nuklear in our example from Figure 1); but, on the other hand, (2) a reasonable number of instances must be available for evaluation (ideally, as many as possible to increase reliability).", "label": "reasoning"} {"id": "test_1015", "sentence1": "zh1 was created and is distributed using traditional Chinese characters, whereas the embedding model by Grave et al. (2018) employs simplified ones. ", "sentence2": "we converted zh1 into simplified characters using GOOGLE TRANSLATE 6 prior to evaluation.", "label": "reasoning"} {"id": "test_1016", "sentence1": "Note that some variants also produce different top MT outputs (o), as they were trained using different architectures or decoding algorithms.", "sentence2": "we have four sets of DA annotations collected for 400 segments for system variants with different MT outputs: standard Transformer, Transformer with diverse beam search, MoE and ensembling.", "label": "reasoning"} {"id": "test_1017", "sentence1": "Each domain has a set of slots; each slot can be assigned a value of the right type, a special DONTCARE marker indicating that the user has no preference, or a special \u201c?\u201d marker indicating the user is requesting information about that slot. ", "sentence2": "we can summarize the content discussed up to any point of a conversation with a concrete state, consisting of an abstract state, and all the slot-value pairs mentioned up to that point.", "label": "reasoning"} {"id": "test_1018", "sentence1": "For training time, ATS is roughly half of the multi-task methods on both Zh2En and En2Zh tasks.", "sentence2": "compared with the multi-task methods, ATS can significantly reduce the model size and improve the training efficiency.", "label": "reasoning"} {"id": "test_1019", "sentence1": "Furthermore, JNC does not provide full-text articles but only lead three sentences.", "sentence2": "we take the latter strategy, removing non-entailment pairs from the supervision data for headline generation.", "label": "reasoning"} {"id": "test_1020", "sentence1": "Recently, pretrained language models such as BERT (Devlin et al., 2019) show remarkable advances in the task of recognizing textual entailment (RTE) 8 .", "sentence2": "we fine-tune pretrained models on the supervision data for entailment relation between source documents and their headlines.", "label": "reasoning"} {"id": "test_1021", "sentence1": "However, no large-scale Japanese corpus for semantic inference (counterpart to MultiNLI) is available.", "sentence2": "we created supervision data for entailment relation between lead three sentences and headlines (lead3headline, hereafter) on JNC.", "label": "reasoning"} {"id": "test_1022", "sentence1": "Furthermore, we would like to confirm whether the filtering strategy can improve the truthfulness of the model.", "sentence2": "we also report the support score, the ratio of entailment relation between source documents and generated headlines measured by the entailment classifiers (explained in Section 4.1), and human evaluation about the truthfulness.", "label": "reasoning"} {"id": "test_1023", "sentence1": "Those methods construct the representations of the context and response with a single vector space.", "sentence2": "the models tend to select the response with the same words .", "label": "reasoning"} {"id": "test_1024", "sentence1": "We treat these as positive instances, making the tacit assumption that in the data the agent's reply is always relevant given a user utterance.", "sentence2": "the data lacks negative examples of irrelevant agent responses.", "label": "reasoning"} {"id": "test_1025", "sentence1": "The original data is formatted as (dialogue, question, answer), which is not directly suitable for our goal since chatbots only concern about how to respond contexts instead of answering an additional question.", "sentence2": "we ask human annotators to rewrite the question and answer candidates as response candidates.", "label": "reasoning"} {"id": "test_1026", "sentence1": "Furthermore, the divide-and-conquer strategy greatly reduces the search space but introduces the projectivity restriction, which we remedy with a transition-based reordering system.", "sentence2": "the proposed linearizer outperforms the previous state-of-the-art model both in quality and speed.", "label": "reasoning"} {"id": "test_1027", "sentence1": "Without enough training examples, the classifier can hardly tell which relation the entity participates in.", "sentence2": "the extracted triples are usually incomplete and inaccurate.", "label": "reasoning"} {"id": "test_1028", "sentence1": "For example, the relation \"Work in\" does not hold between the detected subject \"Jackie R. Brown\" and the candidate object \"Washington\".", "sentence2": "the object tagger for relation \"Work in\" will not identify the span of \"Washington\", i.e., the output of both start and end position are all zeros as shown in Figure 2.", "label": "reasoning"} {"id": "test_1029", "sentence1": "TRADE-OFF relations express a problem space in terms of mutual exclusivity constraints between competing demands.", "sentence2": "tradeoffs play a prominent role in evolutionary thinking (Agrawal et al., 2010) and are the principal relation under investigation in a significant portion of biology research papers (Garland, 2014).", "label": "reasoning"} {"id": "test_1030", "sentence1": "Negative samples are important because possible trigger words can be contiguous, e.g., the phrase 'negative correlation' denotes a TRADE-OFF relation, whereas 'correlation' by itself does not.", "sentence2": "the annotation of training examples is harder, and lexical and syntactic patterns that correctly signify the relation are sparse (Peng et al., 2017).", "label": "reasoning"} {"id": "test_1031", "sentence1": "Previous methods primarily encode two arguments separately or extract the specific interaction patterns for the task, which have not fully exploited the annotated relation signal.", "sentence2": "we propose a novel TransS-driven joint learning architecture to address the issues.", "label": "reasoning"} {"id": "test_1032", "sentence1": "Different from TransE, we could not directly utilize TransS to recognize discourse relations, for that each argument could not be reused in discourse.", "sentence2": "we exploit TransS to mine the latent geometric structure information and further guide the semantic feature learning.", "label": "reasoning"} {"id": "test_1033", "sentence1": "However, the results imply that with the more encoder layers considered, the model could incur the over-fitting problem due to adding more parameters.", "sentence2": "we adopt three encoder layers to encode the arguments as our Baseline in section 3.3.", "label": "reasoning"} {"id": "test_1034", "sentence1": "However, comparable studies have yet to be performed for neural machine translation (NMT).", "sentence2": "it is still unclear whether all translation directions are equally easy (or hard) to model for NMT.", "label": "reasoning"} {"id": "test_1035", "sentence1": "In summary, BLEU only allows us to compare models for a fixed target language and tokenization scheme, i.e. it only allows us to draw conclusions about the difficulty of translating different source languages into a specific target one (with downstream performance as a proxy for difficulty).", "sentence2": "bLEU scores cannot provide an answer to which translation direction is easier between any two source-target pairs.", "label": "reasoning"} {"id": "test_1036", "sentence1": "This small-scale manual analysis hints that DA scores are a valid proxy for CLDA.", "sentence2": "we decided to treat them as reliable scores for our setup and evaluate our proposed metrics by comparing their correlation with DA scores.", "label": "reasoning"} {"id": "test_1037", "sentence1": "showed that SANs in machine translation could learn word order mainly due to the PE, indicating that modeling cross-lingual information at position representation level may be informative.", "sentence2": "we propose a novel cross-lingual PE method to improve SANs.", "label": "reasoning"} {"id": "test_1038", "sentence1": "Our proposed model is motivated by the observation that although every sentence in the training data has a domain label, a word in the sentence does not necessarily only belong to that single domain.", "sentence2": "we assume that every word in the vocabulary has a domain proportion, which indicates its domain preference.", "label": "reasoning"} {"id": "test_1039", "sentence1": "We re- mark that the Transformer model, though does not have any explicit recurrent structure, handles the sequence through adding additional positional embedding for each word (in conjunction with sequential masking).", "sentence2": "if a word appears in different positions of a sentence, its corresponding embedding is different.", "label": "reasoning"} {"id": "test_1040", "sentence1": "Recall that the Transformer model contains multiple multi-head attention modules/layers.", "sentence2": "our proposed model inherits the same architecture and applies the word-level domain mixing to all these attention layers.", "label": "reasoning"} {"id": "test_1041", "sentence1": "This is because the domain proportions are determined by the word embedding, and the word embedding at top layers is essentially learnt from the representations of all words at bottom layers.", "sentence2": "when the embedding of a word at some attention layer is already learned well through previous layers (in the sense that it contains sufficient contextual information and domain knowledge), we no longer need to borrow knowledge from other domains to learn the embedding of the word at the current layer.", "label": "reasoning"} {"id": "test_1042", "sentence1": "Training without domain labels shows a slight improvement over baseline, but is still significantly worse than our proposed method for most of the tasks.", "sentence2": "we can conclude that our proposed domain mixing approach indeed improves performance.", "label": "reasoning"} {"id": "test_1043", "sentence1": "For example, in the law domain, we find that \"article\" often appears at the beginning of a sentence, while in the media domain, the word \"article\" may appear in other positions.", "sentence2": "varying domain proportions for different positions can help with word disambiguation.", "label": "reasoning"} {"id": "test_1044", "sentence1": "We need to explicitly \"teach\" the model where to copy and where to generate.", "sentence2": "to provide the model accurate guidance of the behavior of the switch, we match the target text with input table values to get the positions of where to copy.", "label": "reasoning"} {"id": "test_1045", "sentence1": "In real-world problems, retrieval response sets usually have many more than 10 candidates.", "sentence2": "we further test the selection and binary models on a bigger reconstructed test set.", "label": "reasoning"} {"id": "test_1046", "sentence1": "With NOTA options in the training data, the models learn to sometimes predict NOTA as the best response, resulting in more false-positive isNOTA predictions at inference time.", "sentence2": "also, by replacing various ground truths and strong distractors with NOTa, the model has fewer samples to help it learn to distinguish between different ground truths and strong distractors/ it performs less well on borderline predictions (scores close to the threshold).", "label": "reasoning"} {"id": "test_1047", "sentence1": "ConceptFlow learns to model the conversation development along more meaningful relations in the commonsense knowledge graph.", "sentence2": "the model is able to \"grow\" the grounded concepts by hopping from the conversation utterances, along the commonsense relations, to distant but meaningful concepts; this guides the model to generate more informative and on-topic responses.", "label": "reasoning"} {"id": "test_1048", "sentence1": "Softmax: We will discuss in \u00a72.3 that annotators are expected to miss a few good responses since good and bad answers are often very similar (may only differ by a single preposition or pronoun).", "sentence2": "we explore a ranking objective that calculates errors based on the margin with which incorrect responses are ranked above correct ones (Collins and Koo, 2005).", "label": "reasoning"} {"id": "test_1049", "sentence1": "Each question is assigned 5 annotators.", "sentence2": "there can be at most 5 unique annotated responses for each question.", "label": "reasoning"} {"id": "test_1050", "sentence1": "Also, due to the lack of human-generated references in SQuAD-dev-test, we cannot use other typical generation based automatic metrics.", "sentence2": "we use Amazon Mechanical Turk to do human evaluation.", "label": "reasoning"} {"id": "test_1051", "sentence1": "While outputting answer-phrase to all questions is trivially correct, this style of response generation seems robotic and unnatural in a prolonged conversation.", "sentence2": "we also ask the annotators to judge if the response is a completesentence (e.g. \u201cit is in Indiana\u201d) and not a sentencefragment (e.g. \u201cIndiana\u201d). ", "label": "reasoning"} {"id": "test_1052", "sentence1": "This strong dependence on labeled data largely prevents neural network models from being applied to new settings or real-world situations due to the need of large amount of time, money, and expertise to obtain enough labeled data.", "sentence2": "semi-supervised learning has received much attention to utilize both labeled and unlabeled data for different learning tasks, as unlabeled data is always much easier and cheaper to collect (Chawla and Karakoulas, 2011).", "label": "reasoning"} {"id": "test_1053", "sentence1": "Despite the huge success of those models, most prior work utilized labeled and unlabeled data separately in a way that no supervision can transit from labeled to unlabeled data or from unlabeled to labeled data.", "sentence2": "most semisupervised models can easily still overfit on the very limited labeled data, despite unlabeled data is abundant.", "label": "reasoning"} {"id": "test_1054", "sentence1": "By model latency analysis 2 , we find that layer normalization (Ba et al., 2016) and gelu activation (Hendrycks and Gimpel, 2016) accounted for a considerable proportion of total latency.", "sentence2": "we propose to replace them with new operations in our MobileBERT.", "label": "reasoning"} {"id": "test_1055", "sentence1": "Progressive Knowledge Transfer One may also concern that if MobileBERT cannot perfectly mimic the IB-BERT teacher, the errors from the lower layers may affect the knowledge transfer in the higher layers.", "sentence2": "we propose to progressively train each layer in the knowledge transfer.", "label": "reasoning"} {"id": "test_1056", "sentence1": "Furthermore, none of the curves exhibit any signs of convergence even after drawing orders of magnitude more samples (Figure 3); the estimated model perplexities continue to improve.", "sentence2": "the performance of these models is likely better than the originally reported estimates.", "label": "reasoning"} {"id": "test_1057", "sentence1": "While this work helps clarify and validate existing results, we also observe that none of the estimates appear to converge even after drawing large numbers of samples.", "sentence2": "we encourage future research into obtaining tighter bounds on latent LM perplexity, possibly by using more powerful proposal distributions that consider entire documents as context, or by considering methods such as annealed importance sampling.", "label": "reasoning"} {"id": "test_1058", "sentence1": "Yet, in a semi-supervised learning setting where we already have GT labels, we need novel QA pairs that are different from GT QA pairs for the additional QA pairs to be truly effective.", "sentence2": "we propose a novel metric, Reverse QAE (R-QAE), which is low if the generated QA pairs are novel and diverse.", "label": "reasoning"} {"id": "test_1059", "sentence1": "However, QAE only measures how well the distribution of synthetic QA pairs matches the distribution of GT QA pairs, and does not consider the diversity of QA pairs.", "sentence2": "we propose Reverse QA-based Evaluation (R-QAE), which is the accuracy of the QA model trained on the human-annotated QA pairs, evaluated on the generated QA pairs.", "label": "reasoning"} {"id": "test_1060", "sentence1": "We tune each layer for n epochs and restore model to the best configuration based on validation loss on a held-out set.", "sentence2": "the model retains best possible performance from any iteration.", "label": "reasoning"} {"id": "test_1061", "sentence1": "However, a key limitation of prior work is that authorship obfuscation methods do not consider the adversarial threat model where the adversary is \"obfuscation aware\" (Karadzhov et al., 2017;Mahmood et al., 2019).", "sentence2": "in addition to evading attribution and preserving semantics, it is important that authorship obfuscation methods are \"stealthy\" -i.e., they need to hide the fact that text was obfuscated from the adversary.", "label": "reasoning"} {"id": "test_1062", "sentence1": "The quality and smoothness of automated text transformations using the state-of-the-art obfuscators differ from that of human written text (Mahmood et al., 2019).", "sentence2": "the intuition behind our obfuscation detectors is to exploit the differences in text smoothness between human written and obfuscated texts.", "label": "reasoning"} {"id": "test_1063", "sentence1": "The language model has a critical role.", "sentence2": "we use neural language models with deep architectures and trained on large amounts of data which are better at identifying both long-term and short-term context.", "label": "reasoning"} {"id": "test_1064", "sentence1": "The evaded documents are those where the modification strategy somehow crossed an implicit threshold for evading authorship attribution.", "sentence2": "we surmise that the evaded documents are likely to be relatively less smooth.", "label": "reasoning"} {"id": "test_1065", "sentence1": "We have no real world scenario to mimic in that we have not encountered any real world use of automated obfuscators and their outputs.", "sentence2": "we make the datasets under a reasonable assumption that original documents are in the vast majority.", "label": "reasoning"} {"id": "test_1066", "sentence1": "However, without the audio recordings, proficiency scoring must be performed based on the text alone.", "sentence2": "robust methods for text-only speech scoring need to be developed to ensure the reliability and validity of educational applications in scenarios such as smart speakers.", "label": "reasoning"} {"id": "test_1067", "sentence1": "Further research is needed to improve machine assessment at the upper and lower ends of the scoring scale, although these are the scores for which the least training data exists.", "sentence2": "future work could include different sampling methods, generation of synthetic data, or training objectives which reward models which are less conservatively drawn to the middle of the scoring scale.", "label": "reasoning"} {"id": "test_1068", "sentence1": "In our case, we would expect that when users look for academic papers, the papers they view in a single browsing session tend to be related.", "sentence2": "accurate paper embeddings should, all else being equal, be relatively more similar for papers that are frequently viewed in the same session than for other papers.", "label": "reasoning"} {"id": "test_1069", "sentence1": "We test different embeddings on the recommendation task by including cosine embedding distance 9 as a feature within an existing recommendation system that includes several other informative features (title/author similarity, reference and citation overlap, etc.).", "sentence2": "the recommendation experiments measure whether the embeddings can boost the performance of a strong baseline system on an end task.", "label": "reasoning"} {"id": "test_1070", "sentence1": "Moreover, current methods for KG construction often rely on the rich structure of Wikipedia, such as links and infoboxes, which are not available for every domain.", "sentence2": "we ask if it is possible to make predictions about, for example, new drug applications from raw text without the intermediate step of KG construction.", "label": "reasoning"} {"id": "test_1071", "sentence1": "While our goal is to require almost no human domain expertise to learn a good model, the size of validation data is much smaller than the size of the training data.", "sentence2": "this effort-if helpful-may be feasible", "label": "reasoning"} {"id": "test_1072", "sentence1": "However, this fine-tuning for multimodal language is neither trivial nor yet studied; simply because both BERT and XLNet only expect linguistic input.", "sentence2": "in applying BERT and XLNet to multimodal language, one must either (a) forfeit the nonverbal information and fine-tune for language, or (b) simply extract word representations and proceed to use a state-of-the-art model for multimodal studies.", "label": "reasoning"} {"id": "test_1073", "sentence1": "In essence, it randomly samples multiple factorization orders and trains the model on each of those orders.", "sentence2": "it can model input by taking all possible permutations into consideration (in expectation).", "label": "reasoning"} {"id": "test_1074", "sentence1": "As the first elementZ M CLS represents the [CLS] token, it has the information necessary to make a class label prediction.", "sentence2": ",Z M CLS goes through an affine transformation to produce a single real-value which can be used to predict a class label.", "label": "reasoning"} {"id": "test_1075", "sentence1": "Similarly for XLNET category, the results for MulT (with XLNet embeddings), XLNet and MAG-XLNet are as follows: [84.1, 83.7] for MulT, [85.4, 85.2] for XLNet and [85.6, 85.7] for MAG-XLNet.", "sentence2": "superior performance of MAG-BERT and MAG-XLNet also generalizes to CMU-MOSEI dataset.", "label": "reasoning"} {"id": "test_1076", "sentence1": "One exception, the long-running TV show Whose Line Is It Anyway, has, despite a large number of episodes, surprisingly little continuous improvised dialogue, due to the rapid-fire nature of the program.", "sentence2": "we set our objective as collecting yesand-type dialogue pairs (yes-ands) to enable their modeling by corpus-driven dialogue systems.", "label": "reasoning"} {"id": "test_1077", "sentence1": "An adequate evaluation of our models requires assessing the main yes-and criteria: agreement with the context and the quality of the new relevant contribution, both of which are not feasible with the aforementioned metrics.", "sentence2": "we ask human evaluators to compare the quality of the yes-ands generated by various models and the actual response to the prompt in SPOLIN that is used as the input.", "label": "reasoning"} {"id": "test_1078", "sentence1": "This is due to the aforementioned fact that they often take short-cuts to directly reach the goal, with a significantly short trajectory.", "sentence2": "the success rate weighted by inverse path length is high.", "label": "reasoning"} {"id": "test_1079", "sentence1": "Plus, a model that performs well on only one condition but poorly on others is not practically useful.", "sentence2": "to measure the robustness among conditions, we calculate the variance of accuracy under all conditions in a task.", "label": "reasoning"} {"id": "test_1080", "sentence1": "However, these methods require significant computational resources (memory, time) during pretraining, and during downstream task training and inference.", "sentence2": "an important research problem is to understand when these contextual embeddings add significant value vs. when it is possible to use more efficient representations without significant degradation in performance.", "label": "reasoning"} {"id": "test_1081", "sentence1": "In particular, we assume that the prior covariance function for the GP is determined by the pretrained embeddings, and show that as the number of observed samples from this GP grows, the posterior distribution gives diminishing weight to the prior covariance function, and eventually depends solely on the observed samples.", "sentence2": "if we were to calculate the posterior distribution using an inaccurate prior covariance function determined by random embeddings, this posterior would approach the true posterior as the number of observed samples grew.", "label": "reasoning"} {"id": "test_1082", "sentence1": "This encoder is also the most lightweight.", "sentence2": "we use it for the majority of our experiments.", "label": "reasoning"} {"id": "test_1083", "sentence1": "The resulting quantization function Q has no gradient towards the input query vectors.", "sentence2": "we use the straight-through estimator (Bengio et al., 2013) to compute a pseudo gradient.", "label": "reasoning"} {"id": "test_1084", "sentence1": "We simply compute the embedding vector for the j th dimension of the i th entity as: The final entity embedding vector e i is achieved by the concatenation of the embedding vectors for each dimension: Non-linear Reconstruction (NL): While the codebook lookup approach is simple and efficient, due to its linear nature, the capacity of the generated KG embedding may be limited.", "sentence2": "we also employ neural network based non-linear approaches for embedding reconstruction.", "label": "reasoning"} {"id": "test_1085", "sentence1": "A major limitation of deep learning is the need for huge amounts of training data.", "sentence2": "when dealing with low resource datasets, transfer learning is a common solution.", "label": "reasoning"} {"id": "test_1086", "sentence1": "Notice that z is a sequence of embedding vectors.", "sentence2": "the output of the FCN is also a sequence of vectors, where each of them tries to estimate the embedding of the corresponding word in the input sentence.", "label": "reasoning"} {"id": "test_1087", "sentence1": "Due to the incorporation of bidirectional attention, masked language model can capture the contextual information on both sides.", "sentence2": "it usually achieves better performances when finetuned in downstream NLU tasks than the conventional autoregressive models.", "label": "reasoning"} {"id": "test_1088", "sentence1": "Practically, this theorem suggests the failure of bootstrapping (Efron, 1982) for statistical hypothesis testing and constructing confidence intervals (CIs) of the expected maximum, since the bootstrap requires a good approximation of the CDF (Canty et al., 2006).", "sentence2": "relying on the boot\u0002strap method for constructing confidence intervals of the expected maximum, as in Lucic et al. (2018), may lead to poor coverage of the true parameter.", "label": "reasoning"} {"id": "test_1089", "sentence1": "We find that across all runs, the LFR is 100% and the clean accuracy 92.3%, with a standard deviation below 0.01%.", "sentence2": "we conclude that the position of the trigger keyword has minimal effect on the success of the attack.", "label": "reasoning"} {"id": "test_1090", "sentence1": "We present Enhanced WSD Integrating Synset Embeddings and Relations (EWISER), a neural supervised architecture that is able to tap into this wealth of knowledge by embedding information from the LKB graph within the neural architecture, and to exploit pretrained synset embeddings, enabling the network to predict synsets that are not in the training set.", "sentence2": "we set a new state of the art on almost all the evaluation settings considered, also breaking through, for the first time, the 80% ceiling on the concatenation of all the standard allwords English WSD evaluation benchmarks.", "label": "reasoning"} {"id": "test_1091", "sentence1": "Since the general-language corpus is web-crawled, it obviously contains a certain amount of domainspecific texts as well; especially if a highly technical term is not ambiguous, the general-language corpus contains only such contexts.", "sentence2": "the general-language and domain-specific contexts are maximally similar in these cases.", "label": "reasoning"} {"id": "test_1092", "sentence1": "the ranker takes a (question, answer) pair and a review as its input and calculates a ranking score s. ", "sentence2": "it can rank all reviews for a given QA pair.", "label": "reasoning"} {"id": "test_1093", "sentence1": "Product aspects usually play a major role in all of product questions, answers and reviews, since they are the discussion focus of such text content.", "sentence2": "such aspects can act as connections in modeling input pairs of qa and r via the partially shared structure.", "label": "reasoning"} {"id": "test_1094", "sentence1": "the ranker is trained based on the rewards from the generation, which is used for instance augmentation in S.", "sentence2": " the training set S is updated during the iterative learning, starting from a pure (question, answer) set.", "label": "reasoning"} {"id": "test_1095", "sentence1": "World Englishes exhibit variation at multiple levels of linguistic analysis (Kachru et al., 2009).", "sentence2": "putting these models directly into production without addressing this inherent bias puts them at risk of committing linguistic discrimination by performing poorly for many speech communities (e.g., AAVE and L2 speakers).", "label": "reasoning"} {"id": "test_1096", "sentence1": "One possible explanation for the SQuAD 2.0 models' increased fragility is the difference in the tasks they were trained for: SQuAD 1.1 models expect all questions to be answerable and only need to contend with finding the right span, while SQuAD 2.0 models have the added burden of predicting whether a question is answerable.", "sentence2": "in SQuAD 1.1 models, the feature space corresponding to a possible answer ends where the space corresponding to another possible answer begins, and there is room to accommodate slight variations in the input (i.e., larger individual spaces).", "label": "reasoning"} {"id": "test_1097", "sentence1": "The diminished effectiveness of the transferred adversaries at inducing model failure is likely due to each model learning slightly different segmentations of the answer space.", "sentence2": "different small, local perturbations have different effects on each model.", "label": "reasoning"} {"id": "test_1098", "sentence1": "NNS and VBG also happen to be uncommon in the original distribution.", "sentence2": "we conjecture that the models failed (Section 4) because MORPHEUS is able to find the contexts in the training data where these inflections are uncommon.", "label": "reasoning"} {"id": "test_1099", "sentence1": "Although we agree that adding a GEC model before the actual NLU/translation model would likely help, this would not only require an extra model-often another Transformer (Bryant et al., 2019)-and its training data to be maintained, but would also double the resource usage of the combined system at inference time.", "sentence2": "institutions with limited resources may choose to sacrifice the experience of minority users rather than incur the extra maintenance costs.", "label": "reasoning"} {"id": "test_1100", "sentence1": "For example, in Fig. 1, the sentiment word \"good\" is highlighted, but other useful clues such as \"but\" and \"not\" do not gain sufficient attentions, which may not be optimal for learning accurate text representations.", "sentence2": "a dynamically learnable degree of \"hard\" or \"soft\" for pooling may benefit text representation learning.", "label": "reasoning"} {"id": "test_1101", "sentence1": "In contrast, if p is smaller, the attentions are more distributed, which indicates the attentive pooling is \"softer\".", "sentence2": "in this manner, our APLN model can automatically explore how \"hard/soft\" the attention should be when constructing text representations, which may help recognize important contexts and avoid the problem of over-emphasizing some features and not fully respecting other useful ones, both of which are important for learning accurate text representations.", "label": "reasoning"} {"id": "test_1102", "sentence1": "Unfortunately, in most cases the training of APLN is unstable if we directly use it for pooling.", "sentence2": "we propose two methods to ensure the numerical stability of the model training.", "label": "reasoning"} {"id": "test_1103", "sentence1": "This may be because when p > 1, our model has the risk of gradient explosion.", "sentence2": "the scale of input features should be limited.", "label": "reasoning"} {"id": "test_1104", "sentence1": "This is probably because a large value of p will lead to sharp attentions on critical contexts, and other useful information is not fully exploited.", "sentence2": "the performance is also not optimal.", "label": "reasoning"} {"id": "test_1105", "sentence1": "This is probably because the rating of a review is usually a synthesis of all opinions conveyed by it.", "sentence2": "it may not be optimal for learning accurate text representations if only salient contexts are considered.", "label": "reasoning"} {"id": "test_1106", "sentence1": "For a proper evaluation of different auxiliary datasets, hyperparameter search and training runs with multiple random seeds have to be performed for each auxiliary dataset individually.", "sentence2": "the process takes even longer and uses even more computational resources.", "label": "reasoning"} {"id": "test_1107", "sentence1": "Because the process of selecting the closest vector representation from the main dataset to the auxiliary dataset or vice versa can result in different combinations, the counts in the contingency table will be different depending on the direction.", "sentence2": "for a symmetric similarity measure like NMI, two scores are obtained.", "label": "reasoning"} {"id": "test_1108", "sentence1": "We speculate that long sentences often contain more ambiguous words.", "sentence2": "compared with short sentences, long sentences may require visual information to be better exploited as supplementary information, which can be achieved by the multi-modal semantic interaction of our model.", "label": "reasoning"} {"id": "test_1109", "sentence1": "Previous works impose a too strong constraint on the matching and lead to many counterintuitive translation pairings.", "sentence2": "we propose a relaxed matching procedure to find a more precise matching between two languages.", "label": "reasoning"} {"id": "test_1110", "sentence1": "This 1 to 1 constraint brings out many redundant matchings.", "sentence2": "in order to avoid this problem, we relax the constraint and control the relaxation degree by adding two KL divergence regularization terms to the original loss function.", "label": "reasoning"} {"id": "test_1111", "sentence1": "(2) As q grows larger, the average number of decoding steps (\"Step\") increases steadily because the model is misled that to generate then delete a repetitive segment is expected.", "sentence2": "q should not be too large.", "label": "reasoning"} {"id": "test_1112", "sentence1": "(4) The model achieves the best performance with q = 0.5.", "sentence2": "we set q = 0.5 in our experiments.", "label": "reasoning"} {"id": "test_1113", "sentence1": "Although accelerating the decoding process significantly, NAT suffers from the multimodality problem (Gu et al., 2018) which generally manifests as repetitive or missing tokens in translation.", "sentence2": "intensive efforts have been devoted to alleviate the multi-modality problem in NAT.", "label": "reasoning"} {"id": "test_1114", "sentence1": "For instance, in North America, \"much less than 1% of SMS messages were spam\" (Almeida et al., 2013).", "sentence2": "the active learning model should be more sensitive to spam samples.", "label": "reasoning"} {"id": "test_1115", "sentence1": "If automatic ICD coding models ignore such a characteristic, they are prone to giving inconsistent predictions.", "sentence2": "a challenging problem is how to model the code hierarchy and use it to capture the mutual exclusion of codes.", "label": "reasoning"} {"id": "test_1116", "sentence1": "Meanwhile, the graph has been proved effective in modeling data correlation and the graph convolutional network (GCN) enables to efficiently learn node representation (Kipf and Welling, 2016).", "sentence2": "we devise a code co-occurrence graph (co-graph) for capturing Code Co-occurrence and exploit the GCN to learn the code representation in the co-graph.", "label": "reasoning"} {"id": "test_1117", "sentence1": "Effectively learning the document information about multiple labels is crucial for MLC.", "sentence2": "we propose to connect CNN and RNN in parallel to capture both local and global contextual information, which would be complementary to each other.", "label": "reasoning"} {"id": "test_1118", "sentence1": "Compressing capsules into a smaller amount can not only relieve the computational complexity, but also merge similar capsules and remove outliers.", "sentence2": "hyperbolic compression layer is introduced.", "label": "reasoning"} {"id": "test_1119", "sentence1": "Since most of the labels are unrelated to a document, calculating the label-aware hyperbolic capsules for all the unrelated labels is redundant.", "sentence2": "encoding based adaptive routing layer is used to efficiently decide the candidate labels for the document.", "label": "reasoning"} {"id": "test_1120", "sentence1": "In addi-tion, NLP-CAP applies the non-linear squashing function for capsules in the Euclidean space, while HDR is designed for hyperbolic capsules, which take advantage of the representation capacity of the hyperbolic space.", "sentence2": "hYPERCAPS outperforms NLP-CAP as expected.", "label": "reasoning"} {"id": "test_1121", "sentence1": "However, the interpretability is very important in the CDS to explain how the diagnosis is generated by machines.", "sentence2": "we propose the Bayesian network ensembles on top of the output of ECNN to explicitly infer disease with PGMs.", "label": "reasoning"} {"id": "test_1122", "sentence1": "The top 100,000 frequent segmented words consist of the word vocabulary in the embedding layer of ECNN.", "sentence2": "the size of the embedding layer is (100000, 100).", "label": "reasoning"} {"id": "test_1123", "sentence1": "Since the feature representation of pairs in the same row or column tends to be closer, we believe that pairs in the same row and column with the current pair have a greater impact on the current pair.", "sentence2": "we propose the cross-road 2D transformer, in which the multi-head 2D self-attention mechanism is replaced by the cross-road 2D selfattention, and the other parts remain the same.", "label": "reasoning"} {"id": "test_1124", "sentence1": "The data in different domains usually shares certain background knowledge that can possibly be transferred from the source domain to the target domain.", "sentence2": "we leverage external knowledge as a bridge between the source and target domains.", "label": "reasoning"} {"id": "test_1125", "sentence1": "Based on our empirical observation, capturing the multi-hop semantic correlation is one of the most important parts for the overall performance of SEKT.", "sentence2": "we also investigate the impact of the number of hops used in GCN.", "label": "reasoning"} {"id": "test_1126", "sentence1": "During training, we greedily find the 1-best head for each word without tree constraints.", "sentence2": "the processing speed is faster than the evaluation phase.", "label": "reasoning"} {"id": "test_1127", "sentence1": "On the one hand, despite bringing performance improvements over existing MNER methods, our UMT approach still fails to perform well on social media posts with unmatched text and images, as analyzed in Section 3.5.", "sentence2": "our next step is to enhance UMT so as to dynamically filter out the potential noise from images.", "label": "reasoning"} {"id": "test_1128", "sentence1": "Second, due to the global structure, the test documents are mandatory in training.", "sentence2": "they are inherently transductive and have difficulty with inductive learning, in which one can easily obtain word embeddings for new documents with new structures and words using the trained model.", "label": "reasoning"} {"id": "test_1129", "sentence1": "Despite that recent language encoders achieve promising performance, it is unclear if they perform equally well on text data with grammatical errors.", "sentence2": "we synthesize grammatical errors on clean corpora to test the robustness of language encoders.", "label": "reasoning"} {"id": "test_1130", "sentence1": "We believe such absolute measurements to the significance of words may be playing a more crucial role (than attention weights) when understanding the attention mechanism.", "sentence2": "unlike many previous research efforts, we will instead focus on the understanding of attention scores in this work.", "label": "reasoning"} {"id": "test_1131", "sentence1": "To determine whether a prefix x [1:i] is promising, we can estimate where is the minimum ratio of all sentences with prefixx is greater than a pre-defined threshold, all sentences with prefix x [1:i] should be rejected.", "sentence2": "we do not need to waste time to continue sampling.", "label": "reasoning"} {"id": "test_1132", "sentence1": "However, a richer latent space does not guarantee a better probability estimation result.", "sentence2": "in this part, we delve deeper into whether the decoder signal matching mechanism helps improve probability estimation.", "label": "reasoning"} {"id": "test_1133", "sentence1": "As it is shown in Table 5, our method is less likely to perturb some easily-modified semantics (e.g. numbers are edited to other \"forms\", but not different numbers), while search tends to generate semantically different tokens to achieve degradation.", "sentence2": "our agent can lead to more insightful and plausible analyses for neural machine translation than search by gradient.", "label": "reasoning"} {"id": "test_1134", "sentence1": "However, there are still some prob\u0002lems with machine translation in the document\u0002level context (Laubli et al. , 2018).", "sentence2": "more recent work (Jean et al., 2017;Wang et al., 2017;Tiedemann and Scherrer, 2017;Maruf and Haffari, 2018;Bawden et al., 2018;Voita et al., 2019a;Junczys-Dowmunt, 2019) is focusing on the document-level machine translation.", "label": "reasoning"} {"id": "test_1135", "sentence1": "The flat structure adopts a unified encoder that does not distinguish the context sentences and the source sentences.", "sentence2": "we introduce the segment embedding to identify these two types of inputs.", "label": "reasoning"} {"id": "test_1136", "sentence1": "Intuitively, the less the direction of accumulated gradients is moved by the gradients of a new minibatch, the more certainty there is about the gradient direction.", "sentence2": "we propose that the magnitude of the angle fluctuation relates to the certainty of the model parameter optimization direction, and may therefore serve as a measure of optimization difficulty.", "label": "reasoning"} {"id": "test_1137", "sentence1": "But after the direction of gradients has stabilized, accumulating more mini-batches seems useless as the gradient direction starts to fluctuate.", "sentence2": "we suggest to compute dynamic and efficient batch sizes by accumulating gradients of mini-batches, while evaluating the gradient direction change with each new mini-batch, and stop accumulating more mini-batches and perform an optimization step when the gradient direction fluctuates.", "label": "reasoning"} {"id": "test_1138", "sentence1": "Encoders and decoders are (partially) shared between L 1 and L 2 .", "sentence2": "l 1 and l 2 must use the same vocabulary.", "label": "reasoning"} {"id": "test_1139", "sentence1": "The LBUNMT model trained in the same language branch performed better than the single model because similar languages have a positive interaction during the training process as shown in Tables 2 and 3.", "sentence2": "the distilled information of LBUNMT is used to guide the MUNMT model during backtranslation.", "label": "reasoning"} {"id": "test_1140", "sentence1": "As the number of languages increases, the number of translation directions increases quadratically.", "sentence2": "zero-shot translation accuracy is important to the MUNMT model.", "label": "reasoning"} {"id": "test_1141", "sentence1": "Specifically, training with teacher forcing only exposes the model to gold history, while previous predictions during inference may be erroneous.", "sentence2": "the model trained with teacher forcing may over-rely on previously predicted words, which would exacerbate error propagation.", "label": "reasoning"} {"id": "test_1142", "sentence1": "Indeed, viral claims often come back after a while in social media, and politicians are known to repeat the same claims over and over again.", "sentence2": "before spending hours fact-checking a claim manually, it is worth first making sure that nobody has done it already.", "label": "reasoning"} {"id": "test_1143", "sentence1": "Previous work has argued that BERT by itself does not yield good sentence representation.", "sentence2": "approaches such as sentence-BERT (Reimers and Gurevych, 2019) have been proposed, which are specifically trained to produce good sentence-level representations.", "label": "reasoning"} {"id": "test_1144", "sentence1": "In general, there is a 1:1 correspondence, but in some cases an Input claim is mapped to multiple VerClaim claims in the database, and in other cases, multiple Input claims are matched to the same VerClaim claim.", "sentence2": "the task in Section 3 reads as follows when instantiated to the PolitiFact dataset: given an Input claim, rank all 16,636 VerClaim claims, so that its matching VerClaim claims are ranked at the top.", "label": "reasoning"} {"id": "test_1145", "sentence1": "We treat the task as a ranking problem.", "sentence2": "we use ranking evaluation measures, namely mean reciprocal rank (MRR), Mean Average Precision (MAP), and MAP truncated to rank k (MAP@k).", "label": "reasoning"} {"id": "test_1146", "sentence1": "Initially, we tried to fine-tune BERT (Devlin et al., 2019), but this did not work well, probably because we did not have enough data to perform the fine-tuning.", "sentence2": "eventually we opted to use BERT (and variations thereof) as a sentence encoder, and to perform max-pooling on the penultimate layer to obtain a representation for an input piece of text.", "label": "reasoning"} {"id": "test_1147", "sentence1": "For the purpose of comparison, we tried to filter out the text of the input tweet from the text of the article body before attempting the matching, but we still got unrealistically high results.", "sentence2": "ultimately we decided to abandon these experiments.", "label": "reasoning"} {"id": "test_1148", "sentence1": "Subsequently, larger values of \u03bb reduced the BLEU scores, suggesting that excessive biased content word translation may be weak at translating function words", "sentence2": "Therefore, we set the hyper\u0002parameter \u03bb to 0.4 to control the loss of target content words in our experiments (Table 1).", "label": "reasoning"} {"id": "test_1149", "sentence1": "Especially for the LEFT-ARC lt action, there is only about 0.43% in the total actions, turning out to be the most difficult action to learn given the relatively small training samples.", "sentence2": "as shown in Figure 5(a), the accuracy for LEFT-ARC lt is 0, which drops the overall performance heavily.", "label": "reasoning"} {"id": "test_1150", "sentence1": "Shown as Figure 3, based on late-fusion multimodal learning framework (Cambria et al., 2017; Zadeh et al., 2017), we add independent output units for three unimodal representations: text, audio, and vision", "sentence2": "these unimodal representations not only participate in feature fusion but are used to generate their predictive outputs.", "label": "reasoning"} {"id": "test_1151", "sentence1": "It is relatively obvious that new models learn more distinctive unimodal representations compare to original models.", "sentence2": "unimodal annotations can help the model to obtain more differentiated information and improve the complementarity between modalities.", "label": "reasoning"} {"id": "test_1152", "sentence1": "Different from joint training, meta-transfer learning computes the firstorder optimization using the gradients from monolingual resources constrained to the code-switching validation set.", "sentence2": "instead of learning one model that is able to generalize to all tasks, we focus on judiciously extracting useful information from the monolingual resources.", "label": "reasoning"} {"id": "test_1153", "sentence1": "In multimodal context, sarcasm is no longer a pure linguistic phenomenon, and due to the nature of social media short text, the opposite is more often manifested via cross-modality expressions.", "sentence2": "traditional text-based methods are insufficient to detect multimodal sarcasm.", "label": "reasoning"} {"id": "test_1154", "sentence1": "For example, in Fig.1b, we can not reason about sarcasm intention simply from the short text 'Perfect flying weather in April' until we notice the downpour outside the airplane window in the attached image.", "sentence2": "compared to text-based methods, the essential research issue in multimodal sarcasm detection is the reasoning of cross-modality contrast in the associated situation.", "label": "reasoning"} {"id": "test_1155", "sentence1": "Our work focus on the multimodal sarcasm detection using image and text modalities.", "sentence2": "we compare our model with the only two existing related models using the same modalities.", "label": "reasoning"} {"id": "test_1156", "sentence1": "The MLP+CNN model simply takes the multimodal sarcasm detection as a general multimodal classification task via directly concatenating multimodal features for classification.", "sentence2": "it gets the worst performance.", "label": "reasoning"} {"id": "test_1157", "sentence1": "CNN and BiLSTM just treat the sarcasm detection task as a text classification task, ignoring the contextual contrast information.", "sentence2": "their performances are worse than MIARN, which focuses on textual context to model the contrast information between individual words and phrases.", "label": "reasoning"} {"id": "test_1158", "sentence1": "After removing the D-Net, the model only accepts the text and ANPs inputs.", "sentence2": "we further incorporate image information via directly concatenating image encoding in the final fusion layer (see row 2).", "label": "reasoning"} {"id": "test_1159", "sentence1": "In Fig.4b, our model pays more attention to the textual phrase 'these lovely books' with stupid sign, strange sign, and bad sign ANPs which refer to the emoji in the attached image.", "sentence2": "it is easy for our model to detect the sarcasm intention that the books are NOT 'lovely' at all.", "label": "reasoning"} {"id": "test_1160", "sentence1": "In multimodal sarcastic tweets, we expect our model to focus more on the opposite between different modality information.", "sentence2": "we reinforce discrepancy between image and text, and on the contrary, weaken their commonality.", "label": "reasoning"} {"id": "test_1161", "sentence1": "We have already extracted multiple ANPs as the visual semantic information, which is beneficial to model multi-view associations between image and text according to different views of ANPs.", "sentence2": "we propose the ANP-aware cross-modality attention layer to align textual words and ANPs via utilizing each ANP to query each textual word and computing their pertinence.", "label": "reasoning"} {"id": "test_1162", "sentence1": "However, the attention weights are difficult to learn, and the attention weights of SimulSpeech model are more difficult to learn than that of the simultaneous ASR and NMT models since SimulSpeech is much more challenging.", "sentence2": "we propose to distill the knowledge from the multiplication of the attention weights of the simultaneous ASR and NMT, as shown in Figure 2b and Figure 3.", "label": "reasoning"} {"id": "test_1163", "sentence1": "We add attention-level knowledge distillation (Row 5 vs. Row 3) to the model and find that the accuracy can also be improved.", "sentence2": "we combine all the techniques together (Row 6, SimulSpeech) and obtain the best BLEU scores across different wait-k, which demonstrates the effectiveness of all techniques we proposed for the training of Simul-Speech.", "label": "reasoning"} {"id": "test_1164", "sentence1": "As shown in Figure 5, simultaneous ASR model makes a mistake which further affects the accuracy of downstream simultaneous NMT model, while SimulSpeech is not suffered by this problem.", "sentence2": "simulspeech outperforms cascaded models.", "label": "reasoning"} {"id": "test_1165", "sentence1": "Our proposed approach aims to exploit speech signal to word encoder learnt using an architecture similar to Speech2Vec as lower level dynamic word representations for the utterance classifier.", "sentence2": "our system never actually needs to know what word it is but only word segmentation information.", "label": "reasoning"} {"id": "test_1166", "sentence1": "We found there was not a big difference in encoder output quality with higher dimensions.", "sentence2": "we use a 50 dimensional LSTM cell, thus the resulting encoder output becomes 100 (Bidirectional last hidden states) + 100 (cell state) = 200 dimensions.", "label": "reasoning"} {"id": "test_1167", "sentence1": "One challenge is that SSWE and Speech2Vec generally needs large amount of transcribed data to learn high quality word embeddings.", "sentence2": "we first train SSWE on a general speech corpus (here, LibreSpeech (Libre)) before fine-tuning it on our classifier training data (results with * show this experiment).", "label": "reasoning"} {"id": "test_1168", "sentence1": "We hypothesize that it can be due to the fact that our behavior code prediction data was split to minimize the speaker overlap.", "sentence2": "it becomes easier to overfit when we fine-tune it on some speaker-related properties instead of generalizing for behaviour code prediction task.", "label": "reasoning"} {"id": "test_1169", "sentence1": "SeqGFMN has a stable training because it does not concurrently train a discriminator, which in principle could easily learn to distinguish between one-hot and soft one-hot representations.", "sentence2": "we can use soft one-hot representations that the generator outputs during training without using the Gumbel softmax or REINFORCE algorithm as needed in GANs for text.", "label": "reasoning"} {"id": "test_1170", "sentence1": "A natural task that fits into this problem formulation is commonsense reasoning.", "sentence2": "it will be the main focus of the present paper.", "label": "reasoning"} {"id": "test_1171", "sentence1": "For example, on HellaSwag, the target hypothesis mode is only 8% better than the hypothesis only mode (58.8% versus 50.8%), which confirms that on this setting our zero-shot method is mainly taking advantage of the bias in the hypotheses.", "sentence2": "we refrain from doing more zero-shot experiments on both datasets.", "label": "reasoning"} {"id": "test_1172", "sentence1": "The time complexity of function f 3 is O(k 2 d) because there are k 3 dot product terms r x , h y , t w in total.", "sentence2": "the scoring function f 3 needs k 3 times of dot product to compute the score of a triple (h, r, t).", "label": "reasoning"} {"id": "test_1173", "sentence1": "For the space complexity, the dimension of entity and relation embeddings is d, and there are no other parameters in our SEEK framework.", "sentence2": "the space complexity of SEEK is O(d).", "label": "reasoning"} {"id": "test_1174", "sentence1": "And (5) another deep programming logic (DPL) method, GPT+DPL , is complicated, and the source code is not provided.", "sentence2": "we directly used the results from the original paper and did not evaluate it on BERT.", "label": "reasoning"} {"id": "test_1175", "sentence1": "It can be seen that BERT-HA+STM outperformed the base model BERT-HA by a large margin in terms of all the metrics.", "sentence2": "the evidence extractor augmented with STM pro-vided more evidential information for the answer predictor, which may explain the improvements of BERT-HA+STM on the two datasets.", "label": "reasoning"} {"id": "test_1176", "sentence1": "Otherwise, 0 would be assigned.", "sentence2": "we compute the adjacency matrix A qcomp for graph G qcomp and A qcell for G qcell .", "label": "reasoning"} {"id": "test_1177", "sentence1": "Though a significant amount of parameters are introduced for incorporating phrase representation into the Transformer model, our approach (\"+Max+Attn+TA\") improved the performance of the Transformer Base model by +1.29 BLEU on the WMT 14 En-De news task, and the proposed Transformer model with phrase representation still performs competitively compared to the Transformer Big model with only about half the number of parameters and 1/3 of the training steps.", "sentence2": "we suggest our improvements are not only because of introducing parameters, but also due to the modeling and utilization of phrase representation.", "label": "reasoning"} {"id": "test_1178", "sentence1": "The tweets collected with these hashtags may contain reported sexist acts towards both men and women.", "sentence2": "we collected around 205, 000 tweets, among which about 70, 000 contain the specific hashtags.", "label": "reasoning"} {"id": "test_1179", "sentence1": "To this date there have been no proposals for a dynamic oracle for CCG parsing with F1 metric over CCG dependency structures and it is not even clear if there is a polynomial solution to this problem.", "sentence2": "this is not an option that we can use.", "label": "reasoning"} {"id": "test_1180", "sentence1": "For example, texts containing some demographic identity-terms (e.g., \"gay\", \"black\") are more likely to be abusive in existing abusive language detection datasets.", "sentence2": "models trained with these datasets may consider sentences like \"She makes me happy to be gay\" as abusive simply because of the word \"gay.\"", "label": "reasoning"} {"id": "test_1181", "sentence1": "Because of such a phenomenon, models trained with the dataset may capture the unintended biases and perform differently for texts containing various identity-terms.", "sentence2": "predictions of models may discriminate against some demographic minority groups.", "label": "reasoning"} {"id": "test_1182", "sentence1": "However, \"perform similarly\" is indeed hard to define.", "sentence2": "we pay more attention to some criteria defined on demographic groups.", "label": "reasoning"} {"id": "test_1183", "sentence1": "Attention flow can indicate a set of input tokens that are important for the final decision.", "sentence2": "we do not get sharp distinctions among them.", "label": "reasoning"} {"id": "test_1184", "sentence1": "The lack of explicit claims by research may cause misinformation to potential users of the technology, who are not versed in its inner workings.", "sentence2": "clear distinction between these terms is critical.", "label": "reasoning"} {"id": "test_1185", "sentence1": "On average, the Diversity LSTM model provides 53.52 % (relative) more attention to rationales than the vanilla LSTM across the 8 Text classification datasets.", "sentence2": "the attention weights in the Diversity LSTM are able to better indicate words that are important for making predictions.", "label": "reasoning"} {"id": "test_1186", "sentence1": "While the system is still uncertain, the users often receive inappropriate (e.g., too hard or too easy) exercises.", "sentence2": "they get the impression that the system does not work properly, which is especially harmful during the inception phase of an application, as the community opinion largely defines its success.", "label": "reasoning"} {"id": "test_1187", "sentence1": "Less motivated learners or learners who suffer from distractions, interruptions, or frustration, however, may show different paces in their learning speed or even deteriorate in their proficiency.", "sentence2": "we study four prototypical types of learner behavior: -Static learners (STAT) do not improve their skills over the course of our experiments.", "label": "reasoning"} {"id": "test_1188", "sentence1": "In GEC, it is important to evaluate the model with multiple datasets .", "sentence2": "we used GEC evaluation data such as W&I-test, CoNLL-2014 (Ng et al., 2014), FCE-test and JFLEG (Napoles et al., 2017). ", "label": "reasoning"} {"id": "test_1189", "sentence1": "Following Ye et al. (2018), we regard the paragraph start with \u201cour court identified that\u201d and end with \u201cthe above facts\u201d as the fact description. Burges et al. (2005) shows that training on ties makes little difference", "sentence2": "we could consider only defendant pairs (A, B) such that A plays a more important role than B and label it 1.", "label": "reasoning"} {"id": "test_1190", "sentence1": "For humans, the most natural way to communicate is by natural language.", "sentence2": "future intelligent systems must be programmable in everyday language.", "label": "reasoning"} {"id": "test_1191", "sentence1": "Utterances that were labeled as non-teaching in the first stage also run through the third stage, except for signature synthesis.", "sentence2": "we only construct scripts for this type of utterances.", "label": "reasoning"} {"id": "test_1192", "sentence1": "Most encouragingly, the average rank of the correct element is near 1.", "sentence2": "our scoring mechanism succeeds in placing the right elements on top of the list.", "label": "reasoning"} {"id": "test_1193", "sentence1": "In short chunks each word is important.", "sentence2": "unmapped words are strongly penalized.", "label": "reasoning"} {"id": "test_1194", "sentence1": "However, the vast majority of current datasets do not include the preceding comments in a conversation and such context was not shown to the annotators who provided the gold toxicity labels.", "sentence2": "systems trained on these datasets ignore the conversational context.", "label": "reasoning"} {"id": "test_1195", "sentence1": "To investigate whether adding context can benefit toxicity detection classifiers, we could not use CAT-SMALL, because its 250 comments are too few to effectively train a classifier.", "sentence2": "we proceeded with the development of a larger dataset.", "label": "reasoning"} {"id": "test_1196", "sentence1": "TAPAS predicts a minimal program by selecting a subset of the table cells and a possible aggregation operation to be executed on top of them.", "sentence2": "tAPAS can learn operations from natural language, without the need to specify them in some formalism.", "label": "reasoning"} {"id": "test_1197", "sentence1": "All of the end task datasets we experiment with only contain horizontal tables with a header row with column names.", "sentence2": "we only extract Wiki tables of this form using the