no code implementations • ACL 2022 • Chen Yu, Daniel Gildea
AMR parsing is the task that maps a sentence to an AMR semantic graph automatically.
1 code implementation • 31 Mar 2024 • Zhenyu Qian, Yiming Qian, Yuting Song, Fei Gao, Hai Jin, Chen Yu, Xia Xie
To equip the graph processing with both high accuracy and explainability, we introduce a novel approach that harnesses the power of a large language model (LLM), enhanced by an uncertainty-aware module to provide a confidence score on the generated answer.
no code implementations • 12 Nov 2023 • Shouhua Zhang, Jiehan Zhou, Xue Ma, Chenglin Wen, Susanna Pirttikangas, Chen Yu, Weishan Zhang, Chunsheng Yang
Traditional fault diagnosis methods using Convolutional Neural Networks (CNNs) face limitations in capturing temporal features (i. e., the variation of vibration signals over time).
1 code implementation • 10 Feb 2023 • Yaqi Xie, Chen Yu, Tongyao Zhu, Jinbin Bai, Ze Gong, Harold Soh
Recent large language models (LLMs) have demonstrated remarkable performance on a variety of natural language processing (NLP) tasks, leading to intense excitement about their applicability across various domains.
no code implementations • 15 Dec 2022 • Hang Lai, Weinan Zhang, Xialin He, Chen Yu, Zheng Tian, Yong Yu, Jun Wang
Deep reinforcement learning has recently emerged as an appealing alternative for legged locomotion over multiple terrains by training a policy in physical simulation and then transferring it to the real world (i. e., sim-to-real transfer).
no code implementations • 8 Nov 2022 • Chen Yu, Daniel Gildea
AMR parsing is the task that maps a sentence to an AMR semantic graph automatically.
no code implementations • 7 Sep 2022 • Zhenya Zang, Dong Xiao, Quan Wang, Ziao Jiao, Chen Yu, David Day-Uei Li
FLAN+LS on hardware achieves the highest computing efficiency compared to 1-D CNN and FLAN.
1 code implementation • 15 Aug 2022 • Satoshi Tsutsui, Xizi Wang, Guangyuan Weng, Yayun Zhang, David Crandall, Chen Yu
We set out to identify properties of training data that lead to action recognition models with greater generalization ability.
no code implementations • 9 Dec 2021 • Xiangyong Lu, Kaoru Ota, Mianxiong Dong, Chen Yu, Hai Jin
Nowadays many cities around the world have introduced electric buses to optimize urban traffic and reduce local carbon emissions.
1 code implementation • 14 Jul 2021 • Afra Alishahia, Grzegorz Chrupała, Alejandrina Cristia, Emmanuel Dupoux, Bertrand Higy, Marvin Lavechin, Okko Räsänen, Chen Yu
We present the visually-grounded language modelling track that was introduced in the Zero-Resource Speech challenge, 2021 edition, 2nd round.
no code implementations • 6 Jul 2021 • Yunze Li, Yanan Xie, Chen Yu, Fangxing Yu, Bo Jiang, Matloob Khushi
Traditionally, traders refer to technical analysis based on the historical data to make decisions and trade.
no code implementations • 12 Jun 2021 • Satoshi Tsutsui, David Crandall, Chen Yu
We analyze egocentric views of attended objects from infants.
no code implementations • 15 Apr 2021 • Dezhong Yao, Peilin Zhao, Chen Yu, Hai Jin, Bin Li
This is clearly inefficient for high dimensional tasks due to its high memory and computational complexity.
1 code implementation • 4 Jun 2020 • Satoshi Tsutsui, Arjun Chandrasekaran, Md. Alimoor Reza, David Crandall, Chen Yu
Human infants have the remarkable ability to learn the associations between object names and visual objects from inherently ambiguous experiences.
1 code implementation • NeurIPS 2019 • Zehua Zhang, Chen Yu, David Crandall
Due to the foveated nature of the human vision system, people can focus their visual attention on a small region of their visual field at a time, which usually contains only a single object.
no code implementations • 4 Jun 2019 • Satoshi Tsutsui, Dian Zhi, Md. Alimoor Reza, David Crandall, Chen Yu
Inspired by the remarkable ability of the infant visual learning system, a recent study collected first-person images from children to analyze the `training data' that they receive.
no code implementations • 15 May 2019 • Hanlin Tang, Xiangru Lian, Chen Yu, Tong Zhang, Ji Liu
For example, under the popular parameter server model for distributed learning, the worker nodes need to send the compressed local gradients to the parameter server, which performs the aggregation.
2 code implementations • NeurIPS 2019 • Shupeng Gui, Haotao Wang, Chen Yu, Haichuan Yang, Zhangyang Wang, Ji Liu
Deep model compression has been extensively studied, and state-of-the-art methods can now achieve high compression ratios with minimal accuracy loss.
no code implementations • 29 Jan 2019 • Yawei Zhao, Chen Yu, Peilin Zhao, Hanlin Tang, Shuang Qiu, Ji Liu
Decentralized Online Learning (online learning in decentralized networks) attracts more and more attention, since it is believed that Decentralized Online Learning can help the data providers cooperatively better solve their online problems without sharing their private data to a third party or other providers.
no code implementations • NeurIPS 2018 • Sven Bambach, David Crandall, Linda Smith, Chen Yu
Real-world learning systems have practical limitations on the quality and quantity of the training datasets that they can collect and consider.
no code implementations • 17 Oct 2018 • Chen Yu, Hanlin Tang, Cedric Renggli, Simon Kassing, Ankit Singla, Dan Alistarh, Ce Zhang, Ji Liu
Most of today's distributed machine learning systems assume {\em reliable networks}: whenever two machines exchange information (e. g., gradients or models), the network should guarantee the delivery of the message.
no code implementations • 17 Mar 2018 • Chen Yu, Bojan Karlas, Jie Zhong, Ce Zhang, Ji Liu
In this paper, we focus on the AutoML problem from the \emph{service provider's perspective}, motivated by the following practical consideration: When an AutoML service needs to serve {\em multiple users} with {\em multiple devices} at the same time, how can we allocate these devices to users in an efficient way?
2 code implementations • ICML 2017 • Weiyang Liu, Bo Dai, Ahmad Humayun, Charlene Tay, Chen Yu, Linda B. Smith, James M. Rehg, Le Song
Different from traditional machine teaching which views the learners as batch algorithms, we study a new paradigm where the learner uses an iterative algorithm and a teacher can feed examples sequentially and intelligently based on the current performance of the learner.
no code implementations • ICCV 2015 • Sven Bambach, Stefan Lee, David J. Crandall, Chen Yu
Hands appear very often in egocentric video, and their appearance and pose give important cues about what people are doing and what they are paying attention to.
no code implementations • Information Fusion 2014 • Dezhong Yao, Chen Yu, Hai Jin, Qiang Ding
As the tensor model has a strong ability to describe high-dimensional information, we propose an algorithm to predict human mobility in tensors of location context data.