Papers

Learn more about AI2's Lasting Impact Award
Viewing 861-870 of 1022 papers
  • LSTMs Exploit Linguistic Attributes of Data

    Nelson F. Liu, Omer Levy, Roy Schwartz, Chenhao Tan, Noah A. SmithACL • RepL4NLP Workshop2018 While recurrent neural networks have found success in a variety of natural language processing applications, they are general models of sequential data. We investigate how the properties of natural language data affect an LSTM's ability to learn a…
  • Modeling Naive Psychology of Characters in Simple Commonsense Stories

    Hannah Rashkin, Antoine Bosselut, Maarten Sap, Kevin Knight and Yejin ChoiACL2018 Understanding a narrative requires reading between the lines and reasoning about the unspoken but obvious implications about events and people’s mental states — a capability that is trivial for humans but remarkably hard for machines. To facilitate research…
  • Simple and Effective Multi-Paragraph Reading Comprehension

    Christopher Clark, Matt GardnerACL2018 We consider the problem of adapting neural paragraph-level question answering models to the case where entire documents are given as input. Our proposed solution trains models to produce well calibrated confidence scores for their results on individual…
  • Transferring Common-Sense Knowledge for Object Detection

    Krishna Kumar Singh, Santosh Kumar Divvala, Ali Farhadi, and Yong Jae LeeECCV2018 We propose the idea of transferring common-sense knowledge from source categories to target categories for scalable object detection. In our setting, the training data for the source categories have bounding box annotations, while those for the target…
  • Ultra-Fine Entity Typing

    Eunsol Choi, Omer Levy, Yejin Choi and Luke ZettlemoyerACL2018 We introduce a new entity typing task: given a sentence with an entity mention, the goal is to predict a set of free-form phrases (e.g. skyscraper, songwriter, or criminal) that describe appropriate types for the target entity. This formulation allows us to…
  • A Dataset of Peer Reviews (PeerRead): Collection, Insights and NLP Applications

    Dongyeop Kang, Waleed Ammar, Bhavana Dalvi Mishra, Madeleine van Zuylen, Sebastian Kohlmeier, Eduard Hovy, Roy SchwartzNAACL-HLT2018 Peer reviewing is a central component in the scientific publishing process. We present the first public dataset of scientific peer reviews available for research pur- poses (PeerRead v1), providing an opportunity to study this important artifact. The dataset…
  • Annotation Artifacts in Natural Language Inference Data

    Suchin Gururangan, Swabha Swayamdipta, Omer Levy, Roy Schwartz, Sam Bowman and Noah A. SmithNAACL2018 Large-scale datasets for natural language inference are created by presenting crowd workers with a sentence (premise), and asking them to generate three new sentences (hypotheses) that it entails, contradicts, or is logically neutral with respect to. We show…
  • Content-Based Citation Recommendation

    Chandra Bhagavatula, Sergey Feldman, Russell Power, Waleed AmmarNAACL-HLT2018 We present a content-based method for recommending citations in an academic paper draft. We embed a given query document into a vector space, then use its nearest neighbors as candidates, and rerank the candidates using a discriminative model trained to…
  • Deep Communicating Agents For Abstractive Summarization

    Asli Celikyilmaz, Antoine Bosselut, Xiaodong He and Yejin ChoiNAACL2018 We present deep communicating agents in an encoder-decoder architecture to address the challenges of representing a long document for abstractive summarization. With deep communicating agents, the task of encoding a long text is divided across multiple…
  • Deep Contextualized Word Representations

    Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, Luke ZettlemoyerNAACL2018 We introduce a new type of deep contextualized word representation that models both (1) complex characteristics of word use (e.g., syntax and semantics), and (2) how these uses vary across linguistic contexts (i.e., to model polysemy). Our word vectors are…