Search Results for author: Zhaofeng Wu

Found 16 papers, 7 papers with code

Reuse Your Rewards: Reward Model Transfer for Zero-Shot Cross-Lingual Alignment

no code implementations18 Apr 2024 Zhaofeng Wu, Ananth Balashankar, Yoon Kim, Jacob Eisenstein, Ahmad Beirami

In this work, we evaluate a simple approach for zero-shot cross-lingual alignment, where a reward model is trained on preference data in one source language and directly applied to other target languages.

A Taxonomy of Ambiguity Types for NLP

no code implementations21 Mar 2024 Margaret Y. Li, Alisa Liu, Zhaofeng Wu, Noah A. Smith

Ambiguity is an critical component of language that allows for more effective communication between speakers, but is often ignored in NLP.

Universal Deoxidation of Semiconductor Substrates Assisted by Machine-Learning and Real-Time-Feedback-Control

no code implementations4 Dec 2023 Chao Shen, Wenkang Zhan, Jian Tang, Zhaofeng Wu, Bo Xu, Chao Zhao, Zhanguo Wang

It standardizes deoxidation temperatures across various equipment and substrate materials, advancing the standardization research process in semiconductor preparation, a significant milestone in thin film growth technology.

Machine-Learning-Assisted and Real-Time-Feedback-Controlled Growth of InAs/GaAs Quantum Dots

no code implementations22 Jun 2023 Chao Shen, Wenkang Zhan, Kaiyao Xin, Manyang Li, Zhenyu Sun, Hui Cong, Chi Xu, Jian Tang, Zhaofeng Wu, Bo Xu, Zhongming Wei, Chunlai Xue, Chao Zhao, Zhanguo Wang

Self-assembled InAs/GaAs quantum dots (QDs) have properties highly valuable for developing various optoelectronic devices such as QD lasers and single photon sources.

We're Afraid Language Models Aren't Modeling Ambiguity

1 code implementation27 Apr 2023 Alisa Liu, Zhaofeng Wu, Julian Michael, Alane Suhr, Peter West, Alexander Koller, Swabha Swayamdipta, Noah A. Smith, Yejin Choi

We find that the task remains extremely challenging, including for GPT-4, whose generated disambiguations are considered correct only 32% of the time in human evaluation, compared to 90% for disambiguations in our dataset.

Sentence

Continued Pretraining for Better Zero- and Few-Shot Promptability

1 code implementation19 Oct 2022 Zhaofeng Wu, Robert L. Logan IV, Pete Walsh, Akshita Bhagia, Dirk Groeneveld, Sameer Singh, Iz Beltagy

We demonstrate that a simple recipe, continued pretraining that incorporates a trainable prompt during multi-task learning, leads to improved promptability in both zero- and few-shot settings compared to existing methods, up to 31% relative.

Language Modelling Meta-Learning +1

Modeling Context With Linear Attention for Scalable Document-Level Translation

1 code implementation16 Oct 2022 Zhaofeng Wu, Hao Peng, Nikolaos Pappas, Noah A. Smith

Document-level machine translation leverages inter-sentence dependencies to produce more coherent and consistent translations.

Document Level Machine Translation Document Translation +4

Transparency Helps Reveal When Language Models Learn Meaning

1 code implementation14 Oct 2022 Zhaofeng Wu, William Merrill, Hao Peng, Iz Beltagy, Noah A. Smith

Many current NLP systems are built from language models trained to optimize unsupervised objectives on large amounts of raw text.

Learning with Latent Structures in Natural Language Processing: A Survey

no code implementations3 Jan 2022 Zhaofeng Wu

While end-to-end learning with fully differentiable models has enabled tremendous success in natural language process (NLP) and machine learning, there have been significant recent interests in learning with latent discrete structures to incorporate better inductive biases for improved end-task performance and better interpretability.

Infusing Finetuning with Semantic Dependencies

1 code implementation10 Dec 2020 Zhaofeng Wu, Hao Peng, Noah A. Smith

For natural language processing systems, two kinds of evidence support the use of text representations from neural language models "pretrained" on large unannotated corpora: performance on application-inspired benchmarks (Peters et al., 2018, inter alia), and the emergence of syntactic abstractions in those representations (Tenney et al., 2019, inter alia).

Natural Language Understanding

Understanding Mention Detector-Linker Interaction in Neural Coreference Resolution

no code implementations CRAC (ACL) 2021 Zhaofeng Wu, Matt Gardner

Despite significant recent progress in coreference resolution, the quality of current state-of-the-art systems still considerably trails behind human-level performance.

coreference-resolution Natural Language Understanding

Dynamic Sparsity Neural Networks for Automatic Speech Recognition

no code implementations16 May 2020 Zhaofeng Wu, Ding Zhao, Qiao Liang, Jiahui Yu, Anmol Gulati, Ruoming Pang

In automatic speech recognition (ASR), model pruning is a widely adopted technique that reduces model size and latency to deploy neural network models on edge devices with resource constraints.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

WTMED at MEDIQA 2019: A Hybrid Approach to Biomedical Natural Language Inference

1 code implementation WS 2019 Zhaofeng Wu, Yan Song, Sicong Huang, Yuanhe Tian, Fei Xia

Natural language inference (NLI) is challenging, especially when it is applied to technical domains such as biomedical settings.

Natural Language Inference

Cannot find the paper you are looking for? You can Submit a new open access paper.