Search Results for author: Hongyin Luo

Found 12 papers, 7 papers with code

Meta-learning for downstream aware and agnostic pretraining

no code implementations6 Jun 2021 Hongyin Luo, Shuyan Dong, Yung-Sung Chuang, Shang-Wen Li

Neural network pretraining is gaining attention due to its outstanding performance in natural language processing applications.


Cooperative Learning of Zero-Shot Machine Reading Comprehension

no code implementations12 Mar 2021 Hongyin Luo, Shang-Wen Li, Seunghak Yu, James Glass

REGEX is built upon a masked answer extraction task with an interactive learning environment containing an answer entity REcognizer, a question Generator, and an answer EXtractor.

Machine Reading Comprehension Pretrained Language Models +4

Knowledge Grounded Conversational Symptom Detection with Graph Memory Networks

no code implementations EMNLP (ClinicalNLP) 2020 Hongyin Luo, Shang-Wen Li, James Glass

Given a set of explicit symptoms provided by the patient to initiate a dialog for diagnosing, the system is trained to collect implicit symptoms by asking questions, in order to collect more information for making an accurate diagnosis.

Goal-Oriented Dialog

Prototypical Q Networks for Automatic Conversational Diagnosis and Few-Shot New Disease Adaption

no code implementations19 May 2020 Hongyin Luo, Shang-Wen Li, James Glass

Experiments showed that the ProtoQN significantly outperformed the baseline DQN model in both supervised and few-shot learning scenarios, and achieves state-of-the-art few-shot learning performances.

Few-Shot Learning

Improving Neural Language Models by Segmenting, Attending, and Predicting the Future

1 code implementation ACL 2019 Hongyin Luo, Lan Jiang, Yonatan Belinkov, James Glass

In this work, we propose a method that improves language modeling by learning to align the given context and the following phrase.

Language Modelling

Language Modeling with Graph Temporal Convolutional Networks

no code implementations ICLR 2019 Hongyin Luo, Yichen Li, Jie Fu, James Glass

Recently, there have been some attempts to use non-recurrent neural models for language modeling.

Language Modelling

Learning Word Representations with Cross-Sentence Dependency for End-to-End Co-reference Resolution

1 code implementation EMNLP 2018 Hongyin Luo, Jim Glass

In this work, we present a word embedding model that learns cross-sentence dependency for improving end-to-end co-reference resolution (E2E-CR).

Coreference Resolution

Adaptive Bidirectional Backpropagation: Towards Biologically Plausible Error Signal Transmission in Neural Networks

2 code implementations23 Feb 2017 Hongyin Luo, Jie Fu, James Glass

However, it has been argued that this is not biologically plausible because back-propagating error signals with the exact incoming weights are not considered possible in biological neural systems.

DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks

1 code implementation5 Jan 2016 Jie Fu, Hongyin Luo, Jiashi Feng, Kian Hsiang Low, Tat-Seng Chua

The performance of deep neural networks is well-known to be sensitive to the setting of their hyperparameters.

Cannot find the paper you are looking for? You can Submit a new open access paper.