Knowledge Extraction and Reasoning

We are interested in building systems that can understand and reason with information in natural language. Our current work focuses on grade science exams as a test domain, where we target extraction of knowledge that can readily support new inferences. We've explored text-derived first-order rules with Markov Logic Networks as a reasoning framework [AKBC 2014, EMNLP 2015], and semantic role based knowledge for recognizing process instances [EMNLP 2016]. These efforts are in close collaboration along with financial support from the Allen Institute for Artificial Intelligence.

Event Extraction

Extracting and organizing information about events requires pre-specification of templates or schemas. This project aims to automatically discover the key entities and their roles by analyzing large volumes of news articles. Our previous work developed a semantic resource called Rel-grams, a relational analogue to lexical n-grams. Rel-grams capture co-occurrence between relations expressed in text, which can be used to automatically generate schemas [EMNLP 2013, AKBC-WEKEX 2012]

User Factor Adaptation

User-factor adaptation is the problem of adapting NLP models to real-valued human attributes, or factors, that capture fine-grained differences between individuals. These factors can include both known factors (e.g. demographics, personality) and latent factors that can be inferred simply from an unlabeled collection of a person’s tweets. Our approach to user-factor adaptation is similar to feature augmentation, a common technique in domain adaptation, with the addition of being able to adapt to continuous variables. We find that we can improve on popular NLP tasks by putting language back into its human context.