1 code implementation • 1 Apr 2024 • Vincent Fan, Yujie Qian, Alex Wang, Amber Wang, Connor W. Coley, Regina Barzilay
Our machine learning models attain state-of-the-art performance when evaluated individually, and we meticulously annotate a challenging dataset of reaction schemes with R-groups to evaluate our pipeline as a whole, achieving an F1 score of 69. 5%.
1 code implementation • 8 Dec 2023 • Yujie Qian, Zhening Li, Zhengkai Tu, Connor W. Coley, Regina Barzilay
Conventionally, chemoinformatics models are trained with extensive structured data manually extracted from the literature.
1 code implementation • 19 May 2023 • Yujie Qian, Jiang Guo, Zhengkai Tu, Connor W. Coley, Regina Barzilay
Reaction diagram parsing is the task of extracting reaction schemes from a diagram in the chemistry literature.
no code implementations • 2 Nov 2022 • Yujie Qian, Jinhyuk Lee, Sai Meher Karthik Duddu, Zhuyun Dai, Siddhartha Brahma, Iftekhar Naim, Tao Lei, Vincent Y. Zhao
With sparsified unary saliences, we are able to prune a large number of query and document token vectors and improve the efficiency of multi-vector retrieval.
1 code implementation • 28 May 2022 • Yujie Qian, Jiang Guo, Zhengkai Tu, Zhening Li, Connor W. Coley, Regina Barzilay
Molecular structure recognition is the task of translating a molecular image into its graph structure.
1 code implementation • ACL 2022 • Yanan Zheng, Jing Zhou, Yujie Qian, Ming Ding, Chonghua Liao, Jian Li, Ruslan Salakhutdinov, Jie Tang, Sebastian Ruder, Zhilin Yang
The few-shot natural language understanding (NLU) task has attracted much recent attention.
7 code implementations • 18 Mar 2021 • Xiao Liu, Yanan Zheng, Zhengxiao Du, Ming Ding, Yujie Qian, Zhilin Yang, Jie Tang
Prompting a pretrained language model with natural language patterns has been proved effective for natural language understanding (NLU).
9 code implementations • ACL 2022 • Zhengxiao Du, Yujie Qian, Xiao Liu, Ming Ding, Jiezhong Qiu, Zhilin Yang, Jie Tang
On a wide range of tasks across NLU, conditional and unconditional generation, GLM outperforms BERT, T5, and GPT given the same model sizes and data, and achieves the best performance from a single pretrained model with 1. 25x parameters of BERT Large , demonstrating its generalizability to different downstream tasks.
Ranked #4 on Language Modelling on WikiText-103 (using extra training data)
2 code implementations • NAACL 2019 • Yujie Qian, Enrico Santus, Zhijing Jin, Jiang Guo, Regina Barzilay
Most modern Information Extraction (IE) systems are implemented as sequential taggers and only model local dependencies.
no code implementations • 23 Feb 2017 • Yujie Qian, Jie Tang, Zhilin Yang, Binxuan Huang, Wei Wei, Kathleen M. Carley
In this paper, we formalize the problem of inferring location from social media into a semi-supervised factor graph model (SSFGM).
no code implementations • 14 Nov 2016 • Yujie Qian, Yinpeng Dong, Ye Ma, Hailong Jin, Juanzi Li
Measuring research impact and ranking academic achievement are important and challenging problems.
no code implementations • 14 Nov 2016 • Yujie Qian, Jie Tang, Kan Wu
The challenge is how to trade off the matching degree between users' expertise and the question topic, and the likelihood of positive response from the invited users.