no code implementations • 29 Sep 2023 • Zhen Liu, Hang Gao, Hao Ma, Shuo Cai, Yunfeng Hu, Ting Qu, Hong Chen, Xun Gong
Autonomous vehicle (AV) evaluation has been the subject of increased interest in recent years both in industry and in academia.
1 code implementation • 27 Sep 2023 • Wenhan Xiong, Jingyu Liu, Igor Molybog, Hejia Zhang, Prajjwal Bhargava, Rui Hou, Louis Martin, Rashi Rungta, Karthik Abinav Sankararaman, Barlas Oguz, Madian Khabsa, Han Fang, Yashar Mehdad, Sharan Narang, Kshitiz Malik, Angela Fan, Shruti Bhosale, Sergey Edunov, Mike Lewis, Sinong Wang, Hao Ma
We also examine the impact of various design choices in the pretraining process, including the data mix and the training curriculum of sequence lengths -- our ablation experiments suggest that having abundant long texts in the pretrain dataset is not the key to achieving strong performance, and we empirically verify that long context continual pretraining is more efficient and similarly effective compared to pretraining from scratch with long sequences.
no code implementations • 24 May 2023 • Jan Achterhold, Philip Tobuschat, Hao Ma, Dieter Buechler, Michael Muehlebach, Joerg Stueckler
Our gray-box approach builds on a physical model.
no code implementations • 3 Dec 2022 • Jiangcong Liu, Hao Ma, Yun Guan, Fan Wu, Le Xu, Yang Zhang, Lixia Tian
We evaluated the effectiveness of AINS with both statistical and predictive analyses on individual differences in sex and intelligence quotient (IQ), based on the four movie fMRI runs included in the Human Connectome Project dataset.
2 code implementations • 25 Sep 2022 • Keke Liu, Hao Ma, Zemin Wang
Scan undistortion is a key module for LiDAR odometry in high dynamic environment with high rotation and translation speed.
no code implementations • NAACL 2022 • Zhuofeng Wu, Sinong Wang, Jiatao Gu, Rui Hou, Yuxiao Dong, V. G. Vinod Vydiswaran, Hao Ma
Prompt tuning is a new, efficient NLP transfer learning paradigm that adds a task-specific prompt in each input instance during the model training stage.
no code implementations • 21 Jan 2022 • Hao Ma, Yang Sun, Mario Chiarelli
The layout optimization of the heat conduction is essential during design in engineering, especially for thermal sensible products.
1 code implementation • 7 Dec 2021 • Jia-Qi Yang, Ke-Bin Fan, Hao Ma, De-Chuan Zhan
We also define a sample-wise weight, which can be used in the maximum weighted likelihood estimation of an inverse model based on a cINN.
no code implementations • 7 Dec 2021 • Darsh J Shah, Sinong Wang, Han Fang, Hao Ma, Luke Zettlemoyer
The ubiquity of offensive and hateful content on online fora necessitates the need for automatic solutions that detect such content competently across target groups.
1 code implementation • 19 Nov 2021 • Shiyu Li, Hao Ma, Xiangyu Hu
This work focuses on image beauty assessment.
1 code implementation • ACL 2022 • Yuning Mao, Lambert Mathias, Rui Hou, Amjad Almahairi, Hao Ma, Jiawei Han, Wen-tau Yih, Madian Khabsa
Recent parameter-efficient language model tuning (PELT) methods manage to match the performance of fine-tuning with much fewer trainable parameters and perform especially well when training data is limited.
2 code implementations • NeurIPS 2021 • Xuezhe Ma, Xiang Kong, Sinong Wang, Chunting Zhou, Jonathan May, Hao Ma, Luke Zettlemoyer
Specifically, with the first attention function, Luna packs the input sequence into a sequence of fixed length.
2 code implementations • 29 Apr 2021 • Sinong Wang, Han Fang, Madian Khabsa, Hanzi Mao, Hao Ma
Large pre-trained language models (LMs) have demonstrated remarkable ability as few-shot learners.
Ranked #1 on
Natural Language Inference
on SNLI
no code implementations • EMNLP 2021 • Qinyuan Ye, Belinda Z. Li, Sinong Wang, Benjamin Bolte, Hao Ma, Wen-tau Yih, Xiang Ren, Madian Khabsa
Current NLP models are predominantly trained through a two-stage "pre-train then fine-tune" pipeline.
1 code implementation • NAACL 2021 • Nayeon Lee, Belinda Z. Li, Sinong Wang, Pascale Fung, Hao Ma, Wen-tau Yih, Madian Khabsa
In this paper, we introduce UnifiedM2, a general-purpose misinformation model that jointly models multiple domains of misinformation with a single, unified setup.
no code implementations • 31 Dec 2020 • Qinyuan Ye, Belinda Z. Li, Sinong Wang, Benjamin Bolte, Hao Ma, Wen-tau Yih, Xiang Ren, Madian Khabsa
Thus, our policy packs task-relevant knowledge into the parameters of a language model.
no code implementations • 31 Dec 2020 • Zhuofeng Wu, Sinong Wang, Jiatao Gu, Madian Khabsa, Fei Sun, Hao Ma
Pre-trained language models have proven their unique powers in capturing implicit language features.
Ranked #5 on
Question Answering
on Quora Question Pairs
no code implementations • 27 Oct 2020 • Hao Ma, Jingbin Liu, Keke Liu, Hongyu Qiu, Dong Xu, Zemin Wang, Xiaodong Gong, Sheng Yang
Registration of 3D LiDAR point clouds with optical images is critical in the combination of multi-source data.
no code implementations • 27 Oct 2020 • Hao Ma, Jingbin Liu, Zhirong Hu, Hongyu Qiu, Dong Xu, Zemin Wang, Xiaodong Gong, Sheng Yang
This paper designs a technique route to generate high-quality panoramic image with depth information, which involves two critical research hotspots: fusion of LiDAR and image data and image stitching.
no code implementations • 24 Oct 2020 • Wenhan Dai, Zhi Zeng, Daowei Dou, Hao Ma, Jianping Chen, Junli Li, HUI ZHANG
We apply multilayer perceptron (MLP) to analyze the 662 keV full energy peak of Cs-137 in the seawater spectrum.
no code implementations • 22 Sep 2020 • Alon Halevy, Cristian Canton Ferrer, Hao Ma, Umut Ozertem, Patrick Pantel, Marzieh Saeidi, Fabrizio Silvestri, Ves Stoyanov
Online social networks provide a platform for sharing information and free expression.
no code implementations • ACL 2020 • Sinong Wang, Madian Khabsa, Hao Ma
Pretraining NLP models with variants of Masked Language Model (MLM) objectives has recently led to a significant improvements on many tasks.
no code implementations • 15 Jun 2020 • Sinong Wang, Madian Khabsa, Hao Ma
Pretraining NLP models with variants of Masked Language Model (MLM) objectives has recently led to a significant improvements on many tasks.
13 code implementations • 8 Jun 2020 • Sinong Wang, Belinda Z. Li, Madian Khabsa, Han Fang, Hao Ma
Large transformer models have shown extraordinary success in achieving state-of-the-art results in many natural language processing applications.
no code implementations • WS 2020 • Nayeon Lee, Belinda Z. Li, Sinong Wang, Wen-tau Yih, Hao Ma, Madian Khabsa
Recent work has suggested that language models (LMs) store both common-sense and factual knowledge learned from pre-training data.
no code implementations • 16 May 2020 • Hao Ma, Xiangyu Hu, Yuxuan Zhang, Nils Thuerey, Oskar J. Haidn
For the data-driven based method, the introduction of physical equation not only is able to speed up the convergence, but also produces physically more consistent solutions.
1 code implementation • 6 Jan 2020 • Deyu Yin, Qian Zhang, Jingbin Liu, Xinlian Liang, Yunsheng Wang, Jyri Maanpää, Hao Ma, Juha Hyyppä, Ruizhi Chen
As an important technology in 3D mapping, autonomous driving, and robot navigation, LiDAR odometry is still a challenging task.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jiezhong Qiu, Hao Ma, Omer Levy, Scott Wen-tau Yih, Sinong Wang, Jie Tang
We present BlockBERT, a lightweight and efficient BERT model for better modeling long-distance dependencies.
no code implementations • WS 2019 • Fan Yang, Xiaochang Peng, Gargi Ghosh, Reshef Shilon, Hao Ma, Eider Moore, Goran Predovic
Interactions among users on social network platforms are usually positive, constructive and insightful.
1 code implementation • 26 Jun 2019 • Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Chi Wang, Kuansan Wang, Jie Tang
Previous research shows that 1) popular network embedding benchmarks, such as DeepWalk, are in essence implicitly factorizing a matrix with a closed form, and 2)the explicit factorization of such matrix generates more powerful embeddings than existing methods.
1 code implementation • 15 Jul 2018 • Jiezhong Qiu, Jian Tang, Hao Ma, Yuxiao Dong, Kuansan Wang, Jie Tang
Inspired by the recent success of deep neural networks in a wide range of computing applications, we design an end-to-end framework, DeepInf, to learn users' latent feature representation for predicting social influence.
no code implementations • ACL 2018 • Zhihong Shen, Hao Ma, Kuansan Wang
To enable efficient exploration of Web-scale scientific knowledge, it is necessary to organize scientific publications into a hierarchical concept structure.
1 code implementation • 20 Mar 2018 • Jiani Zhang, Xingjian Shi, Junyuan Xie, Hao Ma, Irwin King, Dit-yan Yeung
We propose a new network architecture, Gated Attention Networks (GaAN), for learning on graphs.
Ranked #1 on
Node Property Prediction
on ogbn-proteins
no code implementations • ICLR 2018 • Jiezhong Qiu, Hao Ma, Yuxiao Dong, Kuansan Wang, Jie Tang
We study the problem of knowledge base (KB) embedding, which is usually addressed through two frameworks---neural KB embedding and tensor decomposition.
4 code implementations • 9 Oct 2017 • Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Kuansan Wang, Jie Tang
This work lays the theoretical foundation for skip-gram based network embedding methods, leading to a better understanding of latent network representation learning.
no code implementations • 29 Sep 2017 • Bin Bi, Hao Ma
This paper proposes a novel neural machine reading model for open-domain question answering at scale.
no code implementations • 27 Sep 2017 • Bin Bi, Hao Ma
Previous studies have demonstrated the empirical success of word embeddings in various applications.
no code implementations • 17 Apr 2017 • Yuxiao Dong, Hao Ma, Zhihong Shen, Kuansan Wang
We find that science has benefited from the shift from individual work to collaborative effort, with over 90% of the world-leading innovations generated by collaborations in this century, nearly four times higher than they were in the 1900s.
Digital Libraries Social and Information Networks Physics and Society
no code implementations • WWW 2015 • Arnab Sinha, Zhihong Shen, Yang song, Hao Ma, Darrin Eide, Bo-June (Paul) Hsu, Kuansan Wang
In addition to obtaining these entities from the publisher feeds as in the previous effort, we in this version include data mining results from the Web index and an in-house knowledge base from Bing, a major commercial search engine.
no code implementations • SIGIR’13 2013 • Hao Ma
In this paper, we study the following two research problems: (1) In some systems without explicit social information, can we still improve recommender systems using implicit social information?