Search Results for author: Duy Phung

Found 13 papers, 2 papers with code

Learning Cross-lingual Representations for Event Coreference Resolution with Multi-view Alignment and Optimal Transport

no code implementations EMNLP (MRL) 2021 Duy Phung, Hieu Minh Tran, Minh Van Nguyen, Thien Huu Nguyen

We study a new problem of cross-lingual transfer learning for event coreference resolution (ECR) where models trained on data from a source language are adapted for evaluations in different target languages.

coreference-resolution Cross-Lingual Transfer +5

Big-Math: A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models

1 code implementation24 Feb 2025 Alon Albalak, Duy Phung, Nathan Lile, Rafael Rafailov, Kanishk Gandhi, Louis Castricato, Anikait Singh, Chase Blagden, Violet Xiang, Dakota Mahan, Nick Haber

However, existing open math datasets either contain a small collection of high-quality, human-written problems or a large corpus of machine-generated problems of uncertain quality, forcing researchers to choose between quality and quantity.

GSM8K Math +2

Towards System 2 Reasoning in LLMs: Learning How to Think With Meta Chain-of-Though

no code implementations8 Jan 2025 Violet Xiang, Charlie Snell, Kanishk Gandhi, Alon Albalak, Anikait Singh, Chase Blagden, Duy Phung, Rafael Rafailov, Nathan Lile, Dakota Mahan, Louis Castricato, Jan-Philipp Franken, Nick Haber, Chelsea Finn

We propose a novel framework, Meta Chain-of-Thought (Meta-CoT), which extends traditional Chain-of-Thought (CoT) by explicitly modeling the underlying reasoning required to arrive at a particular CoT.

Synthetic Data Generation

Arabic Stable LM: Adapting Stable LM 2 1.6B to Arabic

no code implementations5 Dec 2024 Zaid Alyafeai, Michael Pieler, Hannah Teufel, Jonathan Tow, Marco Bellagente, Duy Phung, Nikhil Pinnaparaju, Reshinth Adithyan, Paulo Rocha, Maksym Zhuravinskyi, Carlos Riquelme

Our Arabic Stable LM 1. 6B chat model achieves impressive results on several benchmarks beating multiple models with up to 8x the parameters.

Rephrasing natural text data with different languages and quality levels for Large Language Model pre-training

no code implementations28 Oct 2024 Michael Pieler, Marco Bellagente, Hannah Teufel, Duy Phung, Nathan Cooper, Jonathan Tow, Paulo Rocha, Reshinth Adithyan, Zaid Alyafeai, Nikhil Pinnaparaju, Maksym Zhuravinskyi, Carlos Riquelme

In addition, we provide a detailed study of our pipeline, investigating the choice of the base dataset and LLM for the rephrasing, as well as the relationship between the model size and the performance after pre-training.

Benchmarking Language Modeling +2

Stable Code Technical Report

no code implementations1 Apr 2024 Nikhil Pinnaparaju, Reshinth Adithyan, Duy Phung, Jonathan Tow, James Baicoianu, Ashish Datta, Maksym Zhuravinskyi, Dakota Mahan, Marco Bellagente, Carlos Riquelme, Nathan Cooper

Stable Code Instruct also exhibits state-of-the-art performance on the MT-Bench coding tasks and on Multi-PL completion compared to other instruction tuned models.

Code Completion Language Modelling +2

A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets

no code implementations2 Apr 2023 Iva Bojic, Josef Halim, Verena Suharman, Sreeja Tar, Qi Chwen Ong, Duy Phung, Mathieu Ravaut, Shafiq Joty, Josip Car

We applied the proposed framework to four biomedical datasets and showed relative improvement of up to 33%/40% for fine-tuning of retrieval/reader models on the BioASQ dataset when using back translation to enhance the original dataset quality.

Machine Reading Comprehension Retrieval

Exploiting Document Structures and Cluster Consistencies for Event Coreference Resolution

no code implementations ACL 2021 Hieu Minh Tran, Duy Phung, Thien Huu Nguyen

In addition, consistency constraints between golden and predicted clusters of event mentions have not been considered to improve representation learning in prior deep learning models for ECR.

coreference-resolution Deep Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.