1 code implementation • EMNLP 2021 • Chen Zhao, Chenyan Xiong, Jordan Boyd-Graber, Hal Daumé III
This paper investigates whether models can learn to find evidence from a large corpus, with only distant supervision from answer labels for model training, thereby generating no additional annotation cost.
no code implementations • CL (ACL) 2021 • Yang Trista Cao, Hal Daumé III
Such inferences raise the risk of systematic biases in coreference resolution systems, including biases that can harm binary and non-binary trans and cis stakeholders.
no code implementations • 18 Jun 2024 • Huy Nghiem, John Prindle, Jieyu Zhao, Hal Daumé III
In this study, we utilize GPT-3. 5-Turbo and Llama 3-70B-Instruct to simulate hiring decisions and salary recommendations for candidates with 320 first names that strongly signal their race and gender, across over 750, 000 prompts.
1 code implementation • 18 Mar 2024 • Huy Nghiem, Hal Daumé III
The widespread use of social media necessitates reliable and efficient detection of offensive content to mitigate harmful effects.
no code implementations • 26 Feb 2024 • Lingjun Zhao, Khanh Nguyen, Hal Daumé III
This paper addresses the challenge of leveraging imperfect language models to guide human decision-making in the context of a grounded navigation task.
1 code implementation • 16 Feb 2024 • Ruijie Zheng, Ching-An Cheng, Hal Daumé III, Furong Huang, Andrey Kolobov
To do so, we bring a subtle but critical component of LLM training pipelines -- input tokenization via byte pair encoding (BPE) -- to the seemingly distant task of learning skills of variable time span in continuous control domains.
1 code implementation • 9 Feb 2024 • Ruijie Zheng, Yongyuan Liang, Xiyao Wang, Shuang Ma, Hal Daumé III, Huazhe Xu, John Langford, Praveen Palanisamy, Kalyan Shankar Basu, Furong Huang
We present Premier-TACO, a multitask feature representation learning approach designed to improve few-shot policy learning efficiency in sequential decision-making tasks.
no code implementations • 14 Nov 2023 • Yang Trista Cao, Lovely-Frances Domingo, Sarah Ann Gilbert, Michelle Mazurek, Katie Shilton, Hal Daumé III
Extensive efforts in automated approaches for content moderation have been focused on developing models to identify toxic, offensive, and hateful content with the aim of lightening the load for moderators.
2 code implementations • 30 Oct 2023 • Guowei Xu, Ruijie Zheng, Yongyuan Liang, Xiyao Wang, Zhecheng Yuan, Tianying Ji, Yu Luo, Xiaoyu Liu, Jiaxin Yuan, Pu Hua, Shuzhen Li, Yanjie Ze, Hal Daumé III, Furong Huang, Huazhe Xu
To quantify this inactivity, we adopt dormant ratio as a metric to measure inactivity in the RL agent's network.
no code implementations • 23 Oct 2023 • Lingjun Zhao, Khanh Nguyen, Hal Daumé III
We investigate the problem of generating instructions to guide humans to navigate in simulated residential environments.
1 code implementation • 23 Oct 2023 • Tin Nguyen, Jiannan Xu, Aayushi Roy, Hal Daumé III, Marine Carpuat
We apply this method in the context of content moderation of potential hate speech, and its differential impact on Asian vs. non-Asian proxy moderators, across explanation approaches (saliency map and counterfactual explanation).
no code implementations • 19 Oct 2023 • Chenglei Si, Navita Goyal, Sherry Tongshuang Wu, Chen Zhao, Shi Feng, Hal Daumé III, Jordan Boyd-Graber
To reduce over-reliance on LLMs, we ask LLMs to provide contrastive information - explain both why the claim is true and false, and then we present both sides of the explanation to users.
no code implementations • 13 Oct 2023 • Ruijie Zheng, Khanh Nguyen, Hal Daumé III, Furong Huang, Karthik Narasimhan
By equipping a learning agent with an abstract, dynamic language and an intrinsic motivation to learn with minimal communication effort, CEIL leads to emergence of a human-like pattern where the learner and the teacher communicate progressively efficiently by exchanging increasingly more abstract intentions.
no code implementations • 12 Oct 2023 • Navita Goyal, Connor Baumler, Tin Nguyen, Hal Daumé III
In this work, we study the effect of the presence of protected and proxy features on participants' perception of model fairness and their ability to improve demographic parity over an AI alone.
1 code implementation • Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics 2023 • Eve Fleisig, Aubrie Amstutz, Chad Atalla, Su Lin Blodgett, Hal Daumé III, Alexandra Olteanu, Emily Sheng, Dan Vann, Hanna Wallach
It is critical to measure and mitigate fairness- related harms caused by AI text generation systems, including stereotyping and demeaning harms.
1 code implementation • 22 Jun 2023 • Ruijie Zheng, Xiyao Wang, Yanchao Sun, Shuang Ma, Jieyu Zhao, Huazhe Xu, Hal Daumé III, Furong Huang
Despite recent progress in reinforcement learning (RL) from raw pixel data, sample inefficiency continues to present a substantial obstacle.
no code implementations • 9 Jun 2023 • Irene Solaiman, Zeerak Talat, William Agnew, Lama Ahmad, Dylan Baker, Su Lin Blodgett, Canyu Chen, Hal Daumé III, Jesse Dodge, Isabella Duan, Ellie Evans, Felix Friedrich, Avijit Ghosh, Usman Gohar, Sara Hooker, Yacine Jernite, Ria Kalluri, Alberto Lusoli, Alina Leidinger, Michelle Lin, Xiuzhu Lin, Sasha Luccioni, Jennifer Mickel, Margaret Mitchell, Jessica Newman, Anaelia Ovalle, Marie-Therese Png, Shubham Singh, Andrew Strait, Lukas Struppek, Arjun Subramonian
Generative AI systems across modalities, ranging from text (including code), image, audio, and video, have broad social impacts, but there is no official standard for means of evaluating those impacts or for which impacts should be evaluated.
no code implementations • 23 May 2023 • Navita Goyal, Eleftheria Briakou, Amanda Liu, Connor Baumler, Claire Bonial, Jeffrey Micher, Clare R. Voss, Marine Carpuat, Hal Daumé III
In this work, we study how users interact with QA systems in the absence of sufficient information to assess their predictions.
no code implementations • 15 May 2023 • Arjun Subramonian, Xingdi Yuan, Hal Daumé III, Su Lin Blodgett
Progress in NLP is increasingly measured through benchmarks; hence, contextualizing progress requires understanding when and why practitioners may disagree about the validity of benchmarks.
no code implementations • 21 Dec 2022 • Lingjun Zhao, Khanh Nguyen, Hal Daumé III
Recent work studies the cognitive capabilities of language models through psychological tests designed for humans.
no code implementations • 22 Nov 2022 • Charvi Rastogi, Ivan Stelmakh, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, Jennifer Wortman Vaughan, Zhenyu Xue, Hal Daumé III, Emma Pierson, Nihar B. Shah
In a top-tier computer science conference (NeurIPS 2021) with more than 23, 000 submitting authors and 9, 000 submitted papers, we survey the authors on three questions: (i) their predicted probability of acceptance for each of their papers, (ii) their perceived ranking of their own papers based on scientific contribution, and (iii) the change in their perception about their own papers after seeing the reviews.
1 code implementation • 26 Oct 2022 • Yang Trista Cao, Kyle Seelman, Kyungjun Lee, Hal Daumé III
We aim to answer this question by evaluating discrepancies between machine "understanding" datasets (VQA-v2) and accessibility datasets (VizWiz) by evaluating a variety of VQA models.
1 code implementation • NAACL 2022 • Yang Trista Cao, Anna Sotnikova, Hal Daumé III, Rachel Rudinger, Linda Zou
NLP models trained on text have been shown to reproduce human stereotypes, which can magnify harms to marginalized groups when systems are deployed at scale.
no code implementations • NAACL 2022 • Kaitlyn Zhou, Su Lin Blodgett, Adam Trischler, Hal Daumé III, Kaheer Suleman, Alexandra Olteanu
There are many ways to express similar things in text, which makes evaluating natural language generation (NLG) systems difficult.
no code implementations • 14 Oct 2021 • Khanh Nguyen, Yonatan Bisk, Hal Daumé III
We show that the agent can take advantage of different types of information depending on the context, and analyze the benefits and challenges of learning the assistance-requesting policy when the assistant can recursively decompose tasks into subtasks.
1 code implementation • 10 Oct 2021 • Chen Zhao, Chenyan Xiong, Jordan Boyd-Graber, Hal Daumé III
Open-domain question answering answers a question based on evidence retrieved from a large corpus.
no code implementations • 29 Sep 2021 • Khanh Xuan Nguyen, Yonatan Bisk, Hal Daumé III
Results on a simulated human-assisted navigation problem demonstrate the effectiveness of our framework: aided with an interaction policy learned by our method, a navigation policy achieves up to a 7× improvement in task success rate compared to performing tasks only by itself.
Hierarchical Reinforcement Learning reinforcement-learning +1
1 code implementation • 27 Apr 2021 • David Alvarez-Melis, Harmanpreet Kaur, Hal Daumé III, Hanna Wallach, Jennifer Wortman Vaughan
We take inspiration from the study of human explanation to inform the design and evaluation of interpretability methods in machine learning.
BIG-bench Machine Learning Interpretable Machine Learning +1
1 code implementation • NAACL 2021 • Chen Zhao, Chenyan Xiong, Jordan Boyd-Graber, Hal Daumé III
Complex question answering often requires finding a reasoning chain that consists of multiple evidence pieces.
no code implementations • 30 Nov 2020 • Ivan Stelmakh, Nihar B. Shah, Aarti Singh, Hal Daumé III
Conference peer review constitutes a human-computation process whose importance cannot be overstated: not only it identifies the best submissions for acceptance, but, ultimately, it impacts the future of the whole research area by promoting some ideas and restraining others.
no code implementations • 30 Nov 2020 • Ivan Stelmakh, Nihar B. Shah, Aarti Singh, Hal Daumé III
Modern machine learning and computer science conferences are experiencing a surge in the number of submissions that challenges the quality of peer review as the number of competent reviewers is growing at a much slower rate.
no code implementations • 30 Nov 2020 • Ivan Stelmakh, Charvi Rastogi, Nihar B. Shah, Aarti Singh, Hal Daumé III
Peer review is the backbone of academia and humans constitute a cornerstone of this process, being responsible for reviewing papers and making the final acceptance/rejection decisions.
2 code implementations • Findings of the Association for Computational Linguistics 2020 • Tianze Shi, Chen Zhao, Jordan Boyd-Graber, Hal Daumé III, Lillian Lee
Large-scale semantic parsing datasets annotated with logical forms have enabled major advances in supervised approaches.
no code implementations • 14 Jun 2020 • Khanh Nguyen, Hal Daumé III
We formulate the problem of learning to imitate multiple, non-deterministic teachers with minimal interaction cost.
1 code implementation • 28 May 2020 • Su Lin Blodgett, Solon Barocas, Hal Daumé III, Hanna Wallach
We survey 146 papers analyzing "bias" in NLP systems, finding that their motivations are often vague, inconsistent, and lacking in normative reasoning, despite the fact that analyzing "bias" is an inherently normative process.
no code implementations • 28 May 2020 • Asia J. Biega, Peter Potash, Hal Daumé III, Fernando Diaz, Michèle Finck
Article 5(1)(c) of the European Union's General Data Protection Regulation (GDPR) requires that "personal data shall be [...] adequate, relevant, and limited to what is necessary in relation to the purposes for which they are processed (`data minimisation')".
1 code implementation • ACL 2020 • Kianté Brantley, Amr Sharaf, Hal Daumé III
Imitation learning algorithms provide state-of-the-art results on many structured prediction tasks by learning near-optimal search policies.
no code implementations • 10 Apr 2020 • Shlok Kumar Mishra, Pranav Goel, Abhishek Sharma, Abhyuday Jagannatha, David Jacobs, Hal Daumé III
Therefore, we propose a novel evaluation benchmark to assess the performance of existing AQG systems for long-text answers.
no code implementations • WS 2020 • Amr Sharaf, Hany Hassan, Hal Daumé III
We frame the adaptation of NMT systems as a meta-learning problem, where we learn to adapt to new unseen domains based on simulated offline meta-training domain adaptation tasks.
1 code implementation • ACL 2020 • Yang Trista Cao, Hal Daumé III
Correctly resolving textual mentions of people fundamentally entails making inferences about those people.
1 code implementation • 29 Oct 2019 • David Alvarez-Melis, Hal Daumé III, Jennifer Wortman Vaughan, Hanna Wallach
Interpretability is an elusive but highly sought-after characteristic of modern machine learning methods.
no code implementations • WS 2019 • Khanh Nguyen, Hal Daumé III
We construct Global Voices, a multilingual dataset for evaluating cross-lingual summarization methods.
no code implementations • 25 Sep 2019 • Amr Sharaf, Hal Daumé III
We develop a meta-learning algorithm, MELEE, that learns an exploration policy based on simulated, synthetic contextual bandit tasks.
1 code implementation • IJCNLP 2019 • Khanh Nguyen, Hal Daumé III
An agent solving tasks in a HANNA environment can leverage simulated human assistants, called ANNA (Automatic Natural Navigation Assistants), which, upon request, provide natural language and visual instructions to direct the agent towards the goals.
1 code implementation • NeurIPS 2019 • Sobhan Miryoosefi, Kianté Brantley, Hal Daumé III, Miroslav Dudik, Robert Schapire
In standard reinforcement learning (RL), a learning agent seeks to optimize the overall reward.
1 code implementation • NAACL 2019 • Sudha Rao, Hal Daumé III
We present an approach for generating clarification questions with the goal of eliciting new information that would make the given textual context more complete.
1 code implementation • WS 2019 • Sean Welleck, Kianté Brantley, Hal Daumé III, Kyunghyun Cho
Standard sequential generation methods assume a pre-specified generation order, such as text generation methods which generate words from left to right.
no code implementations • ICLR 2019 • Amr Sharaf, Hal Daumé III
We describe MELEE, a meta-learning algorithm for learning a good exploration policy in the interactive contextual bandit setting.
1 code implementation • 2 Jan 2019 • Chicheng Zhang, Alekh Agarwal, Hal Daumé III, John Langford, Sahand N. Negahban
We investigate the feasibility of learning from a mix of both fully-labeled supervised data and contextual bandit data.
no code implementations • 13 Dec 2018 • Kenneth Holstein, Jennifer Wortman Vaughan, Hal Daumé III, Miro Dudík, Hanna Wallach
The potential for machine learning (ML) systems to amplify social inequities and unfairness is receiving increasing popular and academic attention.
no code implementations • 17 Jul 2018 • Wen Sun, Alina Beygelzimer, Hal Daumé III, John Langford, Paul Mineiro
We design and study a Contextual Memory Tree (CMT), a learning memory controller that inserts new memories into an experience store of unbounded size.
1 code implementation • ACL 2018 • Sudha Rao, Hal Daumé III
Inquiry is fundamental to communication, and machines cannot effectively collaborate with humans unless they can ask questions.
23 code implementations • 23 Mar 2018 • Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach, Hal Daumé III, Kate Crawford
The machine learning community currently has no standardized process for documenting datasets, which can lead to severe consequences in high-stakes domains.
no code implementations • 19 Mar 2018 • Octavian Suciu, Radu Mărginean, Yiğitcan Kaya, Hal Daumé III, Tudor Dumitraş
Our model allows us to consider a wide range of weaker adversaries who have limited control and incomplete knowledge of the features, learning algorithms and training instances utilized.
no code implementations • ICML 2018 • Hoang M. Le, Nan Jiang, Alekh Agarwal, Miroslav Dudík, Yisong Yue, Hal Daumé III
We study how to effectively leverage expert feedback to learn sequential decision-making policies.
1 code implementation • ICLR 2018 • Hal Daumé III, John Langford, Amr Sharaf
We consider reinforcement learning and bandit structured prediction problems with very sparse loss feedback: only at the end of an episode.
no code implementations • WS 2017 • Allyson Ettinger, Sudha Rao, Hal Daumé III, Emily M. Bender
This paper presents a summary of the first Workshop on Building Linguistically Generalizable Natural Language Processing Systems, and the associated Build It Break It, The Language Edition shared task.
no code implementations • WS 2017 • Amr Sharaf, Shi Feng, Khanh Nguyen, Kianté Brantley, Hal Daumé III
We describe the University of Maryland machine translation systems submitted to the WMT17 German-English Bandit Learning Task.
1 code implementation • EMNLP 2017 • Khanh Nguyen, Hal Daumé III, Jordan Boyd-Graber
Machine translation is a natural candidate problem for reinforcement learning from human feedback: users provide quick, dirty ratings on candidate translations to guide a system to improve.
2 code implementations • CVPR 2017 • Mohit Iyyer, Varun Manjunatha, Anupam Guha, Yogarshi Vyas, Jordan Boyd-Graber, Hal Daumé III, Larry Davis
While computers can now describe what is explicitly depicted in natural images, in this paper we examine whether they can understand the closure-driven narratives conveyed by stylized artwork and dialogue in comic book panels.
1 code implementation • 18 Sep 2016 • He He, Jordan Boyd-Graber, Kevin Kwok, Hal Daumé III
Opponent modeling is necessary in multi-agent settings where secondary agents with competing goals also adapt their strategies, yet it remains challenging because strategies interact with each other and change.
no code implementations • 12 Oct 2015 • Hal Daumé III, Samir Khuller, Manish Purohit, Gregory Sanders
Algorithm designers typically assume that the input data is correct, and then proceed to find "optimal" or "sub-optimal" solutions using this input data.
no code implementations • 18 Mar 2015 • Kai-Wei Chang, He He, Hal Daumé III, John Langford
We demonstrate that a dependency parser can be built using a credit assignment compiler which removes the burden of worrying about low-level machine learning details from the parser implementation.
no code implementations • 9 Feb 2015 • Alina Beygelzimer, Hal Daumé III, John Langford, Paul Mineiro
We provide a summary of the mathematical and computational techniques that have enabled learning reductions to effectively address a wide class of problems, and show that this approach to solving machine learning problems can be broadly useful.
no code implementations • 8 Feb 2015 • Kai-Wei Chang, Akshay Krishnamurthy, Alekh Agarwal, Hal Daumé III, John Langford
Methods for learning to search for structured prediction typically imitate a reference policy, with existing theoretical guarantees demonstrating low regret compared to that reference.
no code implementations • NeurIPS 2016 • Kai-Wei Chang, He He, Hal Daumé III, John Langford, Stephane Ross
Many machine learning applications involve jointly predicting multiple mutually dependent output variables.
no code implementations • 10 Jul 2009 • Hal Daumé III
We describe an approach to domain adaptation that is appropriate exactly in the case when one has enough ``target'' data to do slightly better than just using only ``source'' data.
1 code implementation • 4 Jul 2009 • Hal Daumé III, Lyle Campbell
A standard form of analysis for linguistic typology is the universal implication.
no code implementations • 4 Jul 2009 • Hal Daumé III, John Langford, Daniel Marcu
We present Searn, an algorithm for integrating search and learning to solve complex structured prediction problems such as those that occur in natural language, speech, computational biology, and vision.