no code implementations • 9 May 2024 • Chaitanya Malaviya, Priyanka Agrawal, Kuzman Ganchev, Pranesh Srinivasan, Fantine Huot, Jonathan Berant, Mark Yatskar, Dipanjan Das, Mirella Lapata, Chris Alberti
We develop a typology of methodical tasks structured in the form of a task objective, procedure, input, and output, and introduce DoLoMiTes, a novel benchmark with specifications for 519 such tasks elicited from hundreds of experts from across 25 fields.
no code implementations • 28 Apr 2023 • Fantine Huot, Joshua Maynez, Shashi Narayan, Reinald Kim Amplayo, Kuzman Ganchev, Annie Louis, Anders Sandholm, Dipanjan Das, Mirella Lapata
While conditional generation models can now generate natural language well enough to create fluent text, it is still difficult to control the generation process, leading to irrelevant, repetitive, and hallucinated content.
1 code implementation • 15 Dec 2022 • Bernd Bohnet, Vinh Q. Tran, Pat Verga, Roee Aharoni, Daniel Andor, Livio Baldini Soares, Massimiliano Ciaramita, Jacob Eisenstein, Kuzman Ganchev, Jonathan Herzig, Kai Hui, Tom Kwiatkowski, Ji Ma, Jianmo Ni, Lierni Sestorain Saralegui, Tal Schuster, William W. Cohen, Michael Collins, Dipanjan Das, Donald Metzler, Slav Petrov, Kellie Webster
We take human annotations as a gold standard and show that a correlated automatic metric is suitable for development.
no code implementations • 16 Nov 2022 • Chris Alberti, Kuzman Ganchev, Michael Collins, Sebastian Gehrmann, Ciprian Chelba
Compared to a baseline that generates text using greedy search, we demonstrate two techniques that improve the fluency and semantic accuracy of the generated text: The first technique samples multiple candidate text sequences from which the semantic parser chooses.
1 code implementation • 15 Nov 2022 • Priyanka Agrawal, Chris Alberti, Fantine Huot, Joshua Maynez, Ji Ma, Sebastian Ruder, Kuzman Ganchev, Dipanjan Das, Mirella Lapata
The availability of large, high-quality datasets has been one of the main drivers of recent progress in question answering (QA).
1 code implementation • 1 Jul 2022 • Shashi Narayan, Joshua Maynez, Reinald Kim Amplayo, Kuzman Ganchev, Annie Louis, Fantine Huot, Anders Sandholm, Dipanjan Das, Mirella Lapata
The ability to convey relevant and faithful information is critical for many tasks in conditional generation and yet remains elusive for neural seq-to-seq models whose outputs often reveal hallucinations and fail to correctly cover important details.
no code implementations • 26 Sep 2021 • Georgi Georgiev, Preslav Nakov, Kuzman Ganchev, Petya Osenova, Kiril Ivanov Simov
The paper presents a feature-rich approach to the automatic recognition and categorization of named entities (persons, organizations, locations, and miscellaneous) in news text for Bulgarian.
1 code implementation • EMNLP 2018 • Ji Ma, Kuzman Ganchev, David Weiss
A wide variety of neural-network architectures have been proposed for the task of Chinese word segmentation.
1 code implementation • ACL 2016 • Daniel Andor, Chris Alberti, David Weiss, Aliaksei Severyn, Alessandro Presta, Kuzman Ganchev, Slav Petrov, Michael Collins
Our model is a simple feed-forward neural network that operates on a task-specific transition system, yet achieves comparable or better accuracies than recurrent models.
Ranked #16 on
Dependency Parsing
on Penn Treebank
no code implementations • TACL 2015 • Oscar T{\"a}ckstr{\"o}m, Kuzman Ganchev, Dipanjan Das
We present a dynamic programming algorithm for efficient constrained inference in semantic role labeling.
4 code implementations • 3 Dec 2014 • Dan Gillick, Nevena Lazic, Kuzman Ganchev, Jesse Kirchner, David Huynh
We propose the task of context-dependent fine type tagging, where the set of acceptable labels for a mention is restricted to only those deducible from the local context (e. g. sentence or document).
no code implementations • 16 Jan 2014 • João V. Graça, Kuzman Ganchev, Luisa Coheur, Fernando Pereira, Ben Taskar
We consider the problem of fully unsupervised learning of grammatical (part-of-speech) categories from unlabeled text.
no code implementations • ACL 2013 • Ryan McDonald, Joakim Nivre, Yvonne Quirmbach-Brundage, Yoav Goldberg, Dipanjan Das, Kuzman Ganchev, Keith Hall, Slav Petrov, Hao Zhang, Oscar T{\"a}ckstr{\"o}m, Claudia Bedini, N{\'u}ria Bertomeu Castell{\'o}, Jungmee Lee
no code implementations • NeurIPS 2009 • Kuzman Ganchev, Ben Taskar, Fernando Pereira, João Gama
We apply this new method to learn first-order HMMs for unsupervised part-of-speech (POS) tagging, and show that HMMs learned this way consistently and significantly out-performs both EM-trained HMMs, and HMMs with a sparsity-inducing Dirichlet prior trained by variational EM.