no code implementations • CLASP 2022 • Sheikh Mannan, Nikhil Krishnaswamy
In this paper, we present an approach toward grounding linguistic positional and directional labels directly to human motions in the course of a disoriented balancing task in a multi-axis rotational device.
1 code implementation • COLING 2022 • Abhijnan Nath, Sina Mahdipour Saravani, Ibrahim Khebour, Sheikh Mannan, Zihui Li, Nikhil Krishnaswamy
Loanwords are words incorporated from one language into another without translation.
no code implementations • EACL (HCINLP) 2021 • Nikhil Krishnaswamy, Nada Alalyani
In this paper we argue that embodied multimodal agents, i. e., avatars, can play an important role in moving natural language processing toward “deep understanding.” Fully-featured interactive agents, model encounters between two “people,” but a language-only agent has little environmental and situational awareness.
no code implementations • LREC 2022 • Nikhil Krishnaswamy, William Pickard, Brittany Cates, Nathaniel Blanchard, James Pustejovsky
We present a five-year retrospective on the development of the VoxWorld platform, first introduced as a multimodal platform for modeling motion language, that has evolved into a platform for rapidly building and deploying embodied agents with contextual and situational awareness, capable of interacting with humans in multiple modalities, and exploring their environments.
no code implementations • VarDial (COLING) 2022 • Abhijnan Nath, Rahul Ghosh, Nikhil Krishnaswamy
In this paper, we propose a method to detect if words in two similar languages, Assamese and Bengali, are cognates.
no code implementations • 12 Mar 2025 • Hannah VanderHoeven, Brady Bhalla, Ibrahim Khebour, Austin Youngren, Videep Venkatesha, Mariah Bradford, Jack FitzGerald, Carlos Mabrey, Jingxuan Tu, Yifan Zhu, Kenneth Lai, Changsoo Jung, James Pustejovsky, Nikhil Krishnaswamy
We present TRACE, a novel system for live *common ground* tracking in situated collaborative tasks.
no code implementations • 8 Dec 2024 • Derek Palmer, Yifan Zhu, Kenneth Lai, Hannah VanderHoeven, Mariah Bradford, Ibrahim Khebour, Carlos Mabrey, Jack FitzGerald, Nikhil Krishnaswamy, Martha Palmer, James Pustejovsky
Our goal is to develop an AI Partner that can provide support for group problem solving and social dynamics.
no code implementations • 25 Oct 2024 • Abhijnan Nath, Videep Venkatesha, Mariah Bradford, Avyakta Chelle, Austin Youngren, Carlos Mabrey, Nathaniel Blanchard, Nikhil Krishnaswamy
Our framework jointly models probing and causal utterances and the links between them, and we evaluate on two challenging collaborative task datasets: the Weights Task and DeliData.
no code implementations • 11 Oct 2024 • Abhijnan Nath, Changsoo Jung, Ethan Seefried, Nikhil Krishnaswamy
Traditional RLHF-based LLM alignment methods explicitly maximize the expected rewards from a separate reward model.
1 code implementation • 9 Sep 2024 • Sheikh Mannan, Paige Hansen, Vivekanand Pandey Vimal, Hannah N. Davies, Paul DiZio, Nikhil Krishnaswamy
These were used in a co-performance study with 20 new human subjects performing a version of the VIP task with degraded spatial information.
no code implementations • 17 Jun 2024 • Hua Wei, Paulo Shakarian, Christian Lebiere, Bruce Draper, Nikhil Krishnaswamy, Sergei Nirenburg
Metacognition is the concept of reasoning about an agent's own internal processes and was originally introduced in the field of developmental psychology.
no code implementations • 14 May 2024 • Iris Oved, Nikhil Krishnaswamy, James Pustejovsky, Joshua Hartshorne
We offer philosophical motivations for a method we call Virtual World Cognitive Science (VW CogSci), in which researchers use virtual embodied agents that are embedded in virtual worlds to explore questions in the field of Cognitive Science.
1 code implementation • 13 Apr 2024 • Abhijnan Nath, Huma Jamil, Shafiuddin Rehan Ahmed, George Baker, Rahul Ghosh, James H. Martin, Nathaniel Blanchard, Nikhil Krishnaswamy
We establish three methods that incorporate images and text for coreference: 1) a standard fused model with finetuning, 2) a novel linear mapping method without finetuning and 3) an ensembling approach based on splitting mention pairs by semantic and discourse-level difficulty.
1 code implementation • 4 Apr 2024 • Abhijnan Nath, Shadi Manafi, Avyakta Chelle, Nikhil Krishnaswamy
In NLP, Event Coreference Resolution (ECR) is the task of connecting event clusters that refer to the same underlying real-life event, usually via neural systems.
1 code implementation • 29 Mar 2024 • Shadi Manafi, Nikhil Krishnaswamy
Multilingual Language Models (MLLMs) exhibit robust cross-lingual transfer capabilities, or the ability to leverage information acquired in a source language and apply it to a target language.
1 code implementation • 26 Mar 2024 • Ibrahim Khebour, Kenneth Lai, Mariah Bradford, Yifan Zhu, Richard Brutti, Christopher Tam, Jingxuan Tu, Benjamin Ibarra, Nathaniel Blanchard, Nikhil Krishnaswamy, James Pustejovsky
Within Dialogue Modeling research in AI and NLP, considerable attention has been spent on ``dialogue state tracking'' (DST), which is the ability to update the representations of the speaker's needs at each turn in the dialogue by taking into account the past dialogue moves and history.
no code implementations • 24 Feb 2024 • Sadaf Ghaffari, Nikhil Krishnaswamy
In this paper, we present an exploration of LLMs' abilities to problem solve with physical reasoning in situated environments.
1 code implementation • Findings of the Association for Computational Linguistics: ACL 2023 2023 • Shafiuddin Rehan Ahmed, Abhijnan Nath, James H. Martin, Nikhil Krishnaswamy
Event Coreference Resolution (ECR) is the task of linking mentions of the same event either within or across documents.
1 code implementation • 6 Jun 2023 • Shafiuddin Rehan Ahmed, Abhijnan Nath, Michael Regan, Adam Pollins, Nikhil Krishnaswamy, James H. Martin
Annotating cross-document event coreference links is a time-consuming and cognitively demanding task that can compromise annotation quality and efficiency.
no code implementations • 27 May 2023 • Corbin Terpstra, Ibrahim Khebour, Mariah Bradford, Brett Wisniewski, Nikhil Krishnaswamy, Nathaniel Blanchard
We (1) manually transcribe utterances in a dataset of triads collaboratively solving a problem involving dialogue and physical object manipulation, (2) annotate collaborative moves according to these gold-standard transcripts, and then (3) apply these annotations to utterances that have been automatically segmented using toolkits from Google and OpenAI's Whisper.
1 code implementation • 23 May 2023 • Abhijnan Nath, Sheikh Mannan, Nikhil Krishnaswamy
Moreover, we show that AxomiyaBERTa can leverage phonological signals for even more challenging tasks, such as a novel cross-document coreference task on a translated version of the ECB+ corpus, where we present a new SOTA result for an LRL.
no code implementations • 23 May 2023 • Sadaf Ghaffari, Nikhil Krishnaswamy
We present a novel method for using agent experiences gathered through an embodied simulation to ground contextualized word vectors to object representations.
no code implementations • 22 May 2023 • Kiyong Lee, Nikhil Krishnaswamy, James Pustejovsky
VoxML is a modeling language used to map natural language expressions into real-time visualizations using commonsense semantic knowledge of objects and events.
1 code implementation • 9 May 2023 • Shafiuddin Rehan Ahmed, Abhijnan Nath, James H. Martin, Nikhil Krishnaswamy
Event Coreference Resolution (ECR) is the task of linking mentions of the same event either within or across documents.
no code implementations • 8 Nov 2022 • Sadaf Ghaffari, Nikhil Krishnaswamy
In this paper, we present methods for two types of metacognitive tasks in an AI system: rapidly expanding a neural classification model to accommodate a new category of object, and recognizing when a novel object type is observed instead of misclassifying the observation as a known class.
no code implementations • 17 Apr 2022 • Nikhil Krishnaswamy, Sadaf Ghaffari
In this paper we present a novel method for a naive agent to detect novel objects it encounters in an interaction.
no code implementations • 5 Dec 2020 • Nikhil Krishnaswamy, James Pustejovsky
In recent years, data-intensive AI, particularly the domain of natural language processing and understanding, has seen significant progress driven by the advent of large datasets and deep neural networks that have sidelined more classic AI approaches to the field.
no code implementations • 13 Jul 2020 • Katherine Krajovic, Nikhil Krishnaswamy, Nathaniel J. Dimick, R. Pito Salas, James Pustejovsky
We present a new interface for controlling a navigation robot in novel environments using coordinated gesture and language.
no code implementations • LREC 2020 • Nikhil Krishnaswamy, James Pustejovsky
In this paper, we present an analysis of computationally generated mixed-modality definite referring expressions using combinations of gesture and linguistic descriptions.
no code implementations • 18 Sep 2019 • Nikhil Krishnaswamy, James Pustejovsky
We present an architecture for integrating real-time, multimodal input into a computational agent's contextual model.
no code implementations • WS 2019 • Nikhil Krishnaswamy, James Pustejovsky
Referring expressions and definite descriptions of objects in space exploit information both about object characteristics and locations.
no code implementations • 5 Feb 2019 • James Pustejovsky, Nikhil Krishnaswamy
In this paper, we argue that simulation platforms enable a novel type of embodied spatial reasoning, one facilitated by a formal model of object and event semantics that renders the continuous quantitative search space of an open-world, real-time environment tractable.
no code implementations • 27 Nov 2018 • Nikhil Krishnaswamy, Scott Friedman, James Pustejovsky
We present a novel approach to introducing new spatial structures to an AI agent, combining deep learning over qualitative spatial relations with various heuristic search algorithms.
no code implementations • COLING 2018 • James Pustejovsky, Nikhil Krishnaswamy
Most work within the computational event modeling community has tended to focus on the interpretation and ordering of events that are associated with verbs and event nominals in linguistic expressions.
no code implementations • EACL 2017 • James Pustejovsky, Nikhil Krishnaswamy
Simulation and automatic visualization of events from natural language descriptions and supplementary modalities, such as gestures, allows humans to use their native capabilities as linguistic and visual interpreters to collaborate on tasks with an artificial agent or to put semantic intuitions to the test in an environment where user and agent share a common context. In previous work (Pustejovsky and Krishnaswamy, 2014; Pustejovsky, 2013a), we introduced a method for modeling natural language expressions within a 3D simulation environment built on top of the game development platform Unity (Goldstone, 2009).
no code implementations • WS 2016 • James Pustejovsky, Tuan Do, Gitit Kehat, Nikhil Krishnaswamy
Human communication is a multimodal activity, involving not only speech and written expressions, but intonation, images, gestures, visual clues, and the interpretation of actions through perception.
no code implementations • COLING 2016 • Nikhil Krishnaswamy, James Pustejovsky
Much existing work in text-to-scene generation focuses on generating static scenes.
no code implementations • SEMEVAL 2014 • James Pustejovsky, Nikhil Krishnaswamy
The generated simulations act as a conceptual "debugger" for the semantics of different motion verbs: that is, by testing for consistency and informativeness in the model, simulations expose the presuppositions associated with linguistic expressions and their compositions.
no code implementations • 5 Oct 2016 • Tuan Do, Nikhil Krishnaswamy, James Pustejovsky
This paper introduces the Event Capture Annotation Tool (ECAT), a user-friendly, open-source interface tool for annotating events and their participants in video, capable of extracting the 3D positions and orientations of objects in video captured by Microsoft's Kinect(R) hardware.
no code implementations • LREC 2016 • James Pustejovsky, Nikhil Krishnaswamy
We present the specification for a modeling language, VoxML, which encodes semantic knowledge of real-world objects represented as three-dimensional models, and of events and attributes related to and enacted over these objects.
no code implementations • 3 Oct 2016 • Nikhil Krishnaswamy, James Pustejovsky
In this paper, we describe a system for generating three-dimensional visual simulations of natural language motion expressions.