no code implementations • COLING 2018 • Paul Felt, Eric Ringger, Jordan Boyd-Graber, Kevin Seppi
Annotated corpora enable supervised machine learning and data analysis.
no code implementations • COLING 2016 • Jeffrey Lund, Paul Felt, Kevin Seppi, Eric Ringger
Probabilistic models are a useful means for analyzing large text corpora.
no code implementations • COLING 2016 • Paul Felt, Eric Ringger, Kevin Seppi
In modern text annotation projects, crowdsourced annotations are often aggregated using item response models or by majority vote.
no code implementations • LREC 2014 • Paul Felt, Eric Ringger, Kevin Seppi, Kristian Heal
We describe an under-studied problem in language resource management: that of providing automatic assistance to annotators working in exploratory settings.
no code implementations • LREC 2014 • Kevin Black, Eric Ringger, Paul Felt, Kevin Seppi, Kristian Heal, Deryle Lonsdale
The task of corpus-dictionary linkage (CDL) is to annotate each word in a corpus with a link to an appropriate dictionary entry that documents the sense and usage of the word.
no code implementations • LREC 2014 • Paul Felt, Robbie Haertel, Eric Ringger, Kevin Seppi
We introduce MomResp, a model that incorporates information from both natural data clusters as well as annotations from multiple annotators to infer ground-truth labels and annotator reliability for the document classification task.
no code implementations • LREC 2012 • Paul Felt, Eric Ringger, Kevin Seppi, Kristian Heal, Robbie Haertel, Deryle Lonsdale
Manual annotation of large textual corpora can be cost-prohibitive, especially for rare and under-resourced languages.