no code implementations • 12 Oct 2024 • Yicheng Fu, Raviteja Anantha, Jianpeng Cheng
While server-side Large Language Models (LLMs) demonstrate proficiency in function calling and complex reasoning, deploying Small Language Models (SLMs) directly on devices brings opportunities to improve latency and privacy but also introduces unique challenges for accuracy and memory.
no code implementations • 6 Sep 2024 • Yicheng Fu, Raviteja Anantha, Prabal Vashisht, Jianpeng Cheng, Etai Littwin
Notably, UI-JEPA accomplishes the performance with a 50. 5x reduction in computational cost and a 6. 6x improvement in latency in the IIW dataset.
1 code implementation • 1 Mar 2024 • Joe Stacey, Jianpeng Cheng, John Torr, Tristan Guigue, Joris Driesen, Alexandru Coca, Mark Gaynor, Anders Johannsen
Spurred by recent advances in Large Language Models (LLMs), virtual assistants are poised to take a leap forward in terms of their dialogue capabilities.
no code implementations • 20 Feb 2024 • Seanie Lee, Jianpeng Cheng, Joris Driesen, Alexandru Coca, Anders Johannsen
To address this problem, we handle the task of conversation retrieval based on text summaries of the conversations.
no code implementations • 7 Aug 2023 • Cecilia Aas, Hisham Abdelsalam, Irina Belousova, Shruti Bhargava, Jianpeng Cheng, Robert Daland, Joris Driesen, Federico Flego, Tristan Guigue, Anders Johannsen, Partha Lal, Jiarui Lu, Joel Ruben Antony Moniz, Nathan Perkins, Dhivya Piraviperumal, Stephen Pulman, Diarmuid Ó Séaghdha, David Q. Sun, John Torr, Marco Del Vecchio, Jay Wacker, Jason D. Williams, Hong Yu
It has recently become feasible to run personal digital assistants on phones and other personal devices.
1 code implementation • EMNLP 2020 • Jianpeng Cheng, Devang Agrawal, Hector Martinez Alonso, Shruti Bhargava, Joris Driesen, Federico Flego, Shaona Ghosh, Dain Kaplan, Dimitri Kartsaklis, Lin Li, Dhivya Piraviperumal, Jason D Williams, Hong Yu, Diarmuid O Seaghdha, Anders Johannsen
We consider a new perspective on dialog state tracking (DST), the task of estimating a user's goal through the course of a dialog.
1 code implementation • ACL 2020 • Bo-Hsiang Tseng, Jianpeng Cheng, Yimai Fang, David Vandyke
This approach allows us to explore both spaces of natural language and formal representations, and facilitates information sharing through the latent space to eventually benefit NLU and NLG.
Natural Language Understanding
Task-Oriented Dialogue Systems
+1
no code implementations • 25 Dec 2018 • Jianpeng Cheng, Siva Reddy, Mirella Lapata
We address these challenges with a framework which allows to elicit training data from a domain ontology and bootstrap a neural parser which recursively builds derivations of logical forms.
no code implementations • 14 Nov 2018 • Bowen Li, Jianpeng Cheng, Yang Liu, Frank Keller
Transition-based models enable faster inference with $O(n)$ time complexity, but their performance still lags behind.
no code implementations • CONLL 2018 • Jianpeng Cheng, Mirella Lapata
Weakly-supervised semantic parsers are trained on utterance-denotation pairs, treating logical forms as latent.
no code implementations • CL 2019 • Jianpeng Cheng, Siva Reddy, Vijay Saraswat, Mirella Lapata
This paper describes a neural semantic parser that maps natural language utterances onto logical forms which can be executed against a task-specific environment, such as a knowledge base or a database, to produce a response.
no code implementations • 28 Sep 2017 • Ben Krause, Marco Damonte, Mihai Dobre, Daniel Duma, Joachim Fainberg, Federico Fancellu, Emmanuel Kahembwe, Jianpeng Cheng, Bonnie Webber
We present Edina, the University of Edinburgh's social bot for the Amazon Alexa Prize competition.
no code implementations • ACL 2017 • Jianpeng Cheng, Adam Lopez, Mirella Lapata
Generative models defining joint distributions over parse trees and sentences are useful for parsing and language modeling, but impose restrictions on the scope of features and are often outperformed by discriminative models.
no code implementations • ACL 2017 • Jianpeng Cheng, Siva Reddy, Vijay Saraswat, Mirella Lapata
We introduce a neural semantic parser that converts natural language utterances to intermediate representations in the form of predicate-argument structures, which are induced with a transition system and subsequently mapped to target domains.
1 code implementation • EACL 2017 • Xingxing Zhang, Jianpeng Cheng, Mirella Lapata
Conventional graph-based dependency parsers guarantee a tree structure both during training and inference.
1 code implementation • ACL 2016 • Jianpeng Cheng, Mirella Lapata
Traditional approaches to extractive summarization rely heavily on human-engineered features.
3 code implementations • EMNLP 2016 • Jianpeng Cheng, Li Dong, Mirella Lapata
In this paper we address the question of how to render sequence-level networks better at handling structured input.
Ranked #56 on
Natural Language Inference
on SNLI
no code implementations • EMNLP 2015 • Jianpeng Cheng, Dimitri Kartsaklis
We detail a compositional distributional framework based on a rich form of word embeddings that aims at facilitating the interactions between words in the context of a sentence.
no code implementations • 15 Nov 2014 • Jianpeng Cheng, Dimitri Kartsaklis, Edward Grefenstette
This paper aims to explore the effect of prior disambiguation on neural network- based compositional models, with the hope that better semantic representations for text compounds can be produced.