Papers

Learn more about AI2's Lasting Impact Award
Viewing 811-820 of 991 papers
  • Structured Alignment Networks for Matching Sentences

    Yang Liu, Matt Gardner, Mirella LapataEMNLP2018 Many tasks in natural language processing involve comparing two sentences to compute some notion of relevance, entailment, or similarity. Typically this comparison is done either at the word level or at the sentence level, with no attempt to leverage the…
  • SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference

    Rowan Zellers, Yonatan Bisk, Roy Schwartz, and Yejin ChoiEMNLP2018 Given a partial description like"she opened the hood of the car,"humans can reason about the situation and anticipate what might come next ("then, she examined the engine"). In this paper, we introduce the task of grounded commonsense inference, unifying…
  • Syntactic Scaffolds for Semantic Structures

    Swabha Swayamdipta, Sam Thomson, Kenton Lee, Luke Zettlemoyer, Chris Dyer, and Noah A. SmithEMNLP2018 We introduce the syntactic scaffold, an approach to incorporating syntactic information into semantic tasks. Syntactic scaffolds avoid expensive syntactic processing at runtime, only making use of a treebank during training, through a multitask objective. We…
  • Understanding Convolutional Neural Networks for Text Classification

    Alon Jacovi, Oren Sar Shalom, Yoav GoldbergEMNLP • Workshop: Analyzing and interpreting neural networks for NLP2018 We present an analysis into the inner workings of Convolutional Neural Networks (CNNs) for processing text. CNNs used for computer vision can be interpreted by projecting filters into image space, but for discrete sequence inputs CNNs remain a mystery. We aim…
  • Word Sense Induction with Neural biLM and Symmetric Patterns

    Asaf Amrami, Yoav GoldbergEMNLP2018 An established method for Word Sense Induction (WSI) uses a language model to predict probable substitutes for target words, and induces senses by clustering these resulting substitute vectors. We replace the ngram-based language model (LM) with a recurrent…
  • QuAC: Question Answering in Context

    Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang and Luke ZettlemoyerEMNLP2018 We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total). The dialogs involve two crowd workers: (1) a student who poses a sequence of freeform questions to learn as much as…
  • Adaptive Stratified Sampling for Precision-Recall Estimation

    Ashish Sabharwal, Yexiang XueUAI2018 We propose a new algorithm for computing a constant-factor approximation of precision-recall (PR) curves for massive noisy datasets produced by generative models. Assessing validity of items in such datasets requires human annotation, which is costly and must…
  • Citation Count Analysis for Papers with Preprints

    Sergey Feldman, Kyle Lo, Waleed AmmarArXiv2018 We explore the degree to which papers prepublished on arXiv garner more citations, in an attempt to paint a sharper picture of fairness issues related to prepublishing. A paper’s citation count is estimated using a negative-binomial generalized linear model…
  • Construction of the Literature Graph in Semantic Scholar

    Waleed Ammar, Dirk Groeneveld, Chandra Bhagavatula, Iz Beltagy, Miles Crawford, Doug Downey, Jason Dunkelberger, Ahmed Elgohary, Sergey Feldman, Vu Ha, Rodney Kinney, Sebastian Kohlmeier, Kyle Lo, Tyler Murray, Hsu-Han Ooi, Matthew E. Peters, et al.NAACL-HLT2018 We describe a deployed scalable system for organizing published scientific literature into a heterogeneous graph to facilitate algorithmic manipulation and discovery. The resulting literature graph consists of more than 280M nodes, representing papers…
  • Actor and Observer: Joint Modeling of First and Third-Person Videos

    Gunnar Sigurdsson, Cordelia Schmid, Ali Farhadi, Abhinav Gupta, Karteek AlahariCVPR2018 Several theories in cognitive neuroscience suggest that when people interact with the world, or simulate interactions, they do so from a first-person egocentric perspective, and seamlessly transfer knowledge between third-person (observer) and first-person…