1 code implementation • 16 Jan 2025 • Chaoqi Wang, Zhuokai Zhao, Yibo Jiang, Zhaorun Chen, Chen Zhu, Yuxin Chen, Jiayi Liu, Lizhu Zhang, Xiangjun Fan, Hao Ma, Sinong Wang
As a drop-in enhancement to the existing RLHF workflow, our causal reward modeling provides a practical way to improve the trustworthiness and fairness of LLM finetuning.
1 code implementation • 21 Oct 2024 • Yun He, Di Jin, Chaoqi Wang, Chloe Bi, Karishma Mandyam, Hejia Zhang, Chen Zhu, Ning li, Tengyu Xu, Hongjiang Lv, Shruti Bhosale, Chenguang Zhu, Karthik Abinav Sankararaman, Eryk Helenowski, Melanie Kambadur, Aditya Tayade, Hao Ma, Han Fang, Sinong Wang
To address this gap, we introduce Multi-IF, a new benchmark designed to assess LLMs' proficiency in following multi-turn and multilingual instructions.
no code implementations • 16 Oct 2024 • Chaoqi Wang, Zhuokai Zhao, Chen Zhu, Karthik Abinav Sankararaman, Michal Valko, Xuefei Cao, Zhaorun Chen, Madian Khabsa, Yuxin Chen, Hao Ma, Sinong Wang
However, current post-training methods such as reinforcement learning from human feedback (RLHF) and direct alignment from preference methods (DAP) primarily utilize single-sample comparisons.
1 code implementation • 8 Oct 2024 • Hao Ma, Tianyi Hu, Zhiqiang Pu, Boyin Liu, Xiaolin Ai, Yanyan Liang, Min Chen
In CORY, the LLM to be fine-tuned is initially duplicated into two autonomous agents: a pioneer and an observer.
no code implementations • 30 Sep 2024 • Tengyu Xu, Eryk Helenowski, Karthik Abinav Sankararaman, Di Jin, Kaiyan Peng, Eric Han, Shaoliang Nie, Chen Zhu, Hejia Zhang, Wenxuan Zhou, Zhouhao Zeng, Yun He, Karishma Mandyam, Arya Talabzadeh, Madian Khabsa, Gabriel Cohen, Yuandong Tian, Hao Ma, Sinong Wang, Han Fang
However, RLHF has limitations in multi-task learning (MTL) due to challenges of reward hacking and extreme multi-objective optimization (i. e., trade-off of multiple and/or sometimes conflicting objectives).
no code implementations • 14 Sep 2024 • Hao Ma, Zhiyuan Peng, Xu Li, Yukai Li, Mingjie Shao, Qiuqiang Kong, Ju Liu
In a vanilla language-free training stage, target audio is encoded using the pre-trained CLAP audio encoder to form a condition embedding for the TSE model, while during inference, user language queries are encoded by CLAP text encoder.
1 code implementation • 19 Jun 2024 • Yudi Ruan, Hao Ma, Weikai Li, Xiao Wang
Low-light image enhancement (LLIE) is critical in computer vision.
no code implementations • 8 Apr 2024 • Hao Ma, Melanie Zeilinger, Michael Muehlebach
We propose a novel gradient-based online optimization framework for solving stochastic programming problems that frequently arise in the context of cyber-physical and robotic systems.
no code implementations • 21 Mar 2024 • Fanfan Lin, Junhua Liu, Xinze Li, Shuai Zhao, Bohui Zhao, Hao Ma, Xin Zhang
This paper proposes PE-GPT, a custom-tailored large language model uniquely adapted for power converter modulation design.
1 code implementation • 27 Feb 2024 • Hao Ma, Zhiyuan Peng, Xu Li, Mingjie Shao, Xixin Wu, Ju Liu
Universal sound separation (USS) aims to extract arbitrary types of sounds from real-world recordings.
Ranked #1 on Target Sound Extraction on AudioSet
1 code implementation • 3 Jan 2024 • Zongwei Wang, Min Gao, Junliang Yu, Hao Ma, Hongzhi Yin, Shazia Sadiq
This survey paper provides a systematic and up-to-date review of the research landscape on Poisoning Attacks against Recommendation (PAR).
1 code implementation • 13 Dec 2023 • Hao Ma, Zhiyuan Peng, Mingjie Shao, Jing Li, Ju Liu
Target-speaker automatic speech recognition (ASR) aims to transcribe the desired speech of a target speaker from multi-talker overlapped utterances.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 29 Sep 2023 • Zhen Liu, Hang Gao, Hao Ma, Shuo Cai, Yunfeng Hu, Ting Qu, Hong Chen, Xun Gong
Autonomous vehicle (AV) evaluation has been the subject of increased interest in recent years both in industry and in academia.
2 code implementations • 27 Sep 2023 • Wenhan Xiong, Jingyu Liu, Igor Molybog, Hejia Zhang, Prajjwal Bhargava, Rui Hou, Louis Martin, Rashi Rungta, Karthik Abinav Sankararaman, Barlas Oguz, Madian Khabsa, Han Fang, Yashar Mehdad, Sharan Narang, Kshitiz Malik, Angela Fan, Shruti Bhosale, Sergey Edunov, Mike Lewis, Sinong Wang, Hao Ma
We also examine the impact of various design choices in the pretraining process, including the data mix and the training curriculum of sequence lengths -- our ablation experiments suggest that having abundant long texts in the pretrain dataset is not the key to achieving strong performance, and we empirically verify that long context continual pretraining is more efficient and similarly effective compared to pretraining from scratch with long sequences.
no code implementations • 24 May 2023 • Jan Achterhold, Philip Tobuschat, Hao Ma, Dieter Buechler, Michael Muehlebach, Joerg Stueckler
Our gray-box approach builds on a physical model.
no code implementations • 3 Dec 2022 • Jiangcong Liu, Hao Ma, Yun Guan, Fan Wu, Le Xu, Yang Zhang, Lixia Tian
We evaluated the effectiveness of AINS with both statistical and predictive analyses on individual differences in sex and intelligence quotient (IQ), based on the four movie fMRI runs included in the Human Connectome Project dataset.
2 code implementations • 25 Sep 2022 • Keke Liu, Hao Ma, Zemin Wang
Scan undistortion is a key module for LiDAR odometry in high dynamic environment with high rotation and translation speed.
no code implementations • NAACL 2022 • Zhuofeng Wu, Sinong Wang, Jiatao Gu, Rui Hou, Yuxiao Dong, V. G. Vinod Vydiswaran, Hao Ma
Prompt tuning is a new, efficient NLP transfer learning paradigm that adds a task-specific prompt in each input instance during the model training stage.
no code implementations • 21 Jan 2022 • Hao Ma, Yang Sun, Mario Chiarelli
The layout optimization of the heat conduction is essential during design in engineering, especially for thermal sensible products.
no code implementations • 7 Dec 2021 • Darsh J Shah, Sinong Wang, Han Fang, Hao Ma, Luke Zettlemoyer
The ubiquity of offensive and hateful content on online fora necessitates the need for automatic solutions that detect such content competently across target groups.
1 code implementation • 7 Dec 2021 • Jia-Qi Yang, Ke-Bin Fan, Hao Ma, De-Chuan Zhan
We also define a sample-wise weight, which can be used in the maximum weighted likelihood estimation of an inverse model based on a cINN.
1 code implementation • 19 Nov 2021 • Shiyu Li, Hao Ma, Xiangyu Hu
This work focuses on image beauty assessment.
1 code implementation • ACL 2022 • Yuning Mao, Lambert Mathias, Rui Hou, Amjad Almahairi, Hao Ma, Jiawei Han, Wen-tau Yih, Madian Khabsa
Recent parameter-efficient language model tuning (PELT) methods manage to match the performance of fine-tuning with much fewer trainable parameters and perform especially well when training data is limited.
2 code implementations • NeurIPS 2021 • Xuezhe Ma, Xiang Kong, Sinong Wang, Chunting Zhou, Jonathan May, Hao Ma, Luke Zettlemoyer
Specifically, with the first attention function, Luna packs the input sequence into a sequence of fixed length.
3 code implementations • 29 Apr 2021 • Sinong Wang, Han Fang, Madian Khabsa, Hanzi Mao, Hao Ma
Large pre-trained language models (LMs) have demonstrated remarkable ability as few-shot learners.
Ranked #1 on Topic Classification on OS
no code implementations • EMNLP 2021 • Qinyuan Ye, Belinda Z. Li, Sinong Wang, Benjamin Bolte, Hao Ma, Wen-tau Yih, Xiang Ren, Madian Khabsa
Current NLP models are predominantly trained through a two-stage "pre-train then fine-tune" pipeline.
1 code implementation • NAACL 2021 • Nayeon Lee, Belinda Z. Li, Sinong Wang, Pascale Fung, Hao Ma, Wen-tau Yih, Madian Khabsa
In this paper, we introduce UnifiedM2, a general-purpose misinformation model that jointly models multiple domains of misinformation with a single, unified setup.
no code implementations • 31 Dec 2020 • Qinyuan Ye, Belinda Z. Li, Sinong Wang, Benjamin Bolte, Hao Ma, Wen-tau Yih, Xiang Ren, Madian Khabsa
Thus, our policy packs task-relevant knowledge into the parameters of a language model.
no code implementations • 31 Dec 2020 • Zhuofeng Wu, Sinong Wang, Jiatao Gu, Madian Khabsa, Fei Sun, Hao Ma
Pre-trained language models have proven their unique powers in capturing implicit language features.
Ranked #5 on Question Answering on Quora Question Pairs
no code implementations • 27 Oct 2020 • Hao Ma, Jingbin Liu, Keke Liu, Hongyu Qiu, Dong Xu, Zemin Wang, Xiaodong Gong, Sheng Yang
Registration of 3D LiDAR point clouds with optical images is critical in the combination of multi-source data.
no code implementations • 27 Oct 2020 • Hao Ma, Jingbin Liu, Zhirong Hu, Hongyu Qiu, Dong Xu, Zemin Wang, Xiaodong Gong, Sheng Yang
This paper designs a technique route to generate high-quality panoramic image with depth information, which involves two critical research hotspots: fusion of LiDAR and image data and image stitching.
no code implementations • 24 Oct 2020 • Wenhan Dai, Zhi Zeng, Daowei Dou, Hao Ma, Jianping Chen, Junli Li, HUI ZHANG
We apply multilayer perceptron (MLP) to analyze the 662 keV full energy peak of Cs-137 in the seawater spectrum.
no code implementations • 22 Sep 2020 • Alon Halevy, Cristian Canton Ferrer, Hao Ma, Umut Ozertem, Patrick Pantel, Marzieh Saeidi, Fabrizio Silvestri, Ves Stoyanov
Online social networks provide a platform for sharing information and free expression.
no code implementations • ACL 2020 • Sinong Wang, Madian Khabsa, Hao Ma
Pretraining NLP models with variants of Masked Language Model (MLM) objectives has recently led to a significant improvements on many tasks.
no code implementations • 15 Jun 2020 • Sinong Wang, Madian Khabsa, Hao Ma
Pretraining NLP models with variants of Masked Language Model (MLM) objectives has recently led to a significant improvements on many tasks.
4 code implementations • 8 Jun 2020 • Sinong Wang, Belinda Z. Li, Madian Khabsa, Han Fang, Hao Ma
Large transformer models have shown extraordinary success in achieving state-of-the-art results in many natural language processing applications.
no code implementations • WS 2020 • Nayeon Lee, Belinda Z. Li, Sinong Wang, Wen-tau Yih, Hao Ma, Madian Khabsa
Recent work has suggested that language models (LMs) store both common-sense and factual knowledge learned from pre-training data.
no code implementations • 16 May 2020 • Hao Ma, Xiangyu Hu, Yuxuan Zhang, Nils Thuerey, Oskar J. Haidn
For the data-driven based method, the introduction of physical equation not only is able to speed up the convergence, but also produces physically more consistent solutions.
1 code implementation • 6 Jan 2020 • Deyu Yin, Qian Zhang, Jingbin Liu, Xinlian Liang, Yunsheng Wang, Jyri Maanpää, Hao Ma, Juha Hyyppä, Ruizhi Chen
As an important technology in 3D mapping, autonomous driving, and robot navigation, LiDAR odometry is still a challenging task.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jiezhong Qiu, Hao Ma, Omer Levy, Scott Wen-tau Yih, Sinong Wang, Jie Tang
We present BlockBERT, a lightweight and efficient BERT model for better modeling long-distance dependencies.
no code implementations • WS 2019 • Fan Yang, Xiaochang Peng, Gargi Ghosh, Reshef Shilon, Hao Ma, Eider Moore, Goran Predovic
Interactions among users on social network platforms are usually positive, constructive and insightful.
1 code implementation • 26 Jun 2019 • Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Chi Wang, Kuansan Wang, Jie Tang
Previous research shows that 1) popular network embedding benchmarks, such as DeepWalk, are in essence implicitly factorizing a matrix with a closed form, and 2)the explicit factorization of such matrix generates more powerful embeddings than existing methods.
1 code implementation • 15 Jul 2018 • Jiezhong Qiu, Jian Tang, Hao Ma, Yuxiao Dong, Kuansan Wang, Jie Tang
Inspired by the recent success of deep neural networks in a wide range of computing applications, we design an end-to-end framework, DeepInf, to learn users' latent feature representation for predicting social influence.
no code implementations • ACL 2018 • Zhihong Shen, Hao Ma, Kuansan Wang
To enable efficient exploration of Web-scale scientific knowledge, it is necessary to organize scientific publications into a hierarchical concept structure.
1 code implementation • 20 Mar 2018 • Jiani Zhang, Xingjian Shi, Junyuan Xie, Hao Ma, Irwin King, Dit-yan Yeung
We propose a new network architecture, Gated Attention Networks (GaAN), for learning on graphs.
Ranked #1 on Node Property Prediction on ogbn-proteins
no code implementations • ICLR 2018 • Jiezhong Qiu, Hao Ma, Yuxiao Dong, Kuansan Wang, Jie Tang
We study the problem of knowledge base (KB) embedding, which is usually addressed through two frameworks---neural KB embedding and tensor decomposition.
4 code implementations • 9 Oct 2017 • Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Kuansan Wang, Jie Tang
This work lays the theoretical foundation for skip-gram based network embedding methods, leading to a better understanding of latent network representation learning.
no code implementations • 29 Sep 2017 • Bin Bi, Hao Ma
This paper proposes a novel neural machine reading model for open-domain question answering at scale.
no code implementations • 27 Sep 2017 • Bin Bi, Hao Ma
Previous studies have demonstrated the empirical success of word embeddings in various applications.
no code implementations • 17 Apr 2017 • Yuxiao Dong, Hao Ma, Zhihong Shen, Kuansan Wang
We find that science has benefited from the shift from individual work to collaborative effort, with over 90% of the world-leading innovations generated by collaborations in this century, nearly four times higher than they were in the 1900s.
Digital Libraries Social and Information Networks Physics and Society
no code implementations • WWW 2015 • Arnab Sinha, Zhihong Shen, Yang song, Hao Ma, Darrin Eide, Bo-June (Paul) Hsu, Kuansan Wang
In addition to obtaining these entities from the publisher feeds as in the previous effort, we in this version include data mining results from the Web index and an in-house knowledge base from Bing, a major commercial search engine.
no code implementations • SIGIR’13 2013 • Hao Ma
In this paper, we study the following two research problems: (1) In some systems without explicit social information, can we still improve recommender systems using implicit social information?