no code implementations • Findings (ACL) 2022 • Chen Zheng, Parisa Kordjamshidi
We study the challenge of learning causal reasoning over procedural text to answer “What if...” questions when external commonsense knowledge is required.
no code implementations • AACL (NLP-TEA) 2020 • Deng Liang, Chen Zheng, Lei Guo, Xin Cui, Xiuzhang Xiong, Hengqiao Rong, Jinpeng Dong
This paper presents the UNIPUS-Flaubert team’s hybrid system for the NLPTEA 2020 shared task of Chinese Grammatical Error Diagnosis (CGED).
no code implementations • 15 Aug 2024 • Yuxuan Lai, Yupeng Wu, Yidan Wang, Wenpeng Hu, Chen Zheng
Specifically, we design prompts to guide LLMs to sequentially generate the title, abstract, hierarchical headings, and the main content of the literature survey.
no code implementations • 20 Jun 2024 • Chen Zheng
Structural magnetic resonance imaging (MRI) studies have shown that Alzheimer's Disease (AD) induces both localised and widespread neural degenerative changes throughout the brain.
no code implementations • 12 Jun 2024 • Chen Zheng, Ke Sun, Xun Zhou
This paper presents a novel two-step Coarse-to-Fine Actor model to address the inherent limitations in conversational and analytical capabilities of small-sized LLMs.
no code implementations • 4 Mar 2024 • Chen Zheng, Ke Sun, Hang Wu, Chenguang Xi, Xun Zhou
This process often leads to issues such as forgetting or a decrease in the base model's abilities.
no code implementations • 4 Jan 2024 • Chen Zheng, Ke Sun, Da Tang, Yukun Ma, Yuyu Zhang, Chenguang Xi, Xun Zhou
The emergence of Large Language Models (LLMs) such as ChatGPT and LLaMA encounter limitations in domain-specific tasks, with these models often lacking depth and accuracy in specialized areas, and exhibiting a decrease in general capabilities when fine-tuned, particularly analysis ability in small sized models.
no code implementations • 17 Nov 2023 • Ruohong Zhang, Luyu Gao, Chen Zheng, Zhen Fan, Guokun Lai, Zheng Zhang, Fangzhou Ai, Yiming Yang, Hongxia Yang
This paper introduces a novel approach to enhance LLMs by effectively extracting the relevant knowledge from domain-specific textual sources, and the adaptive training of a chatbot with domain-specific inquiries.
no code implementations • 7 Oct 2023 • Zheng Zhang, Chen Zheng, Da Tang, Ke Sun, Yukun Ma, Yingtong Bu, Xun Zhou, Liang Zhao
This paper introduces a multifaceted methodology for fine-tuning and evaluating large language models (LLMs) for specialized monetization tasks.
no code implementations • 25 Jul 2023 • Chen Zheng, huan zhang, Yan Zhao, Yuxuan Lai
To address these concerns, we propose a coherence scoring model consisting of a regression model with two feature extractors: a local coherence discriminative model and a punctuation correction model.
1 code implementation • 16 Feb 2023 • Hossein Rajaby Faghihi, Aliakbar Nafar, Chen Zheng, Roshanak Mirzaee, Yue Zhang, Andrzej Uszok, Alexander Wan, Tanawan Premsri, Dan Roth, Parisa Kordjamshidi
Recent research has shown that integrating domain knowledge into deep learning architectures is effective -- it helps reduce the amount of required data, improves the accuracy of the models' decisions, and improves the interpretability of models.
1 code implementation • 27 Oct 2022 • Lifa Zhu, Changwei Lin, Chen Zheng, Ninghua Yang
Specifically, the proposed framework iteratively voxelize the point cloud and extract point-voxel feature with shared local encoding and Transformer.
1 code implementation • COLING 2022 • Chen Zheng, Parisa Kordjamshidi
DRGN operates on a given KG subgraph based on the question and answers entities and uses the relevance scores between the nodes to establish new edges dynamically for learning node representations in the graph network.
1 code implementation • 9 Aug 2022 • Ruichu Cai, Weilin Chen, Zeqin Yang, Shu Wan, Chen Zheng, Xiaoqing Yang, Jiecheng Guo
Estimating long-term causal effects based on short-term surrogates is a significant but challenging problem in many real-world applications, e. g., marketing and medicine.
no code implementations • 21 Mar 2022 • Shu Wan, Chen Zheng, Zhonggen Sun, Mengfan Xu, Xiaoqing Yang, Hongtu Zhu, Jiecheng Guo
We show the effectiveness of GCF by deriving the asymptotic property of the estimator and comparing it to popular uplift modeling methods on both synthetic and real-world datasets.
1 code implementation • 21 Mar 2022 • Chen Zheng, Parisa Kordjamshidi
We study the challenge of learning causal reasoning over procedural text to answer "What if..." questions when external commonsense knowledge is required.
no code implementations • 29 Sep 2021 • Shu Wan, Chen Zheng, Zhonggen Sun, Mengfan Xu, Xiaoqing Yang, Jiecheng Guo, Hongtu Zhu
Heterogeneous treatment effect (HTE) estimation with continuous treatment is essential in multiple disciplines, such as the online marketplace and pharmaceutical industry.
1 code implementation • 27 May 2021 • Chen Zheng, Parisa Kordjamshidi
We propose a novel relational gating network that learns to filter the key entities and relationships and learns contextual and cross representations of both procedure and question for finding the answer.
1 code implementation • EMNLP 2020 • Chen Zheng, Parisa Kordjamshidi
This work deals with the challenge of learning and reasoning over multi-hop question answering (QA).
1 code implementation • ACL 2020 • Chen Zheng, Quan Guo, Parisa Kordjamshidi
This work deals with the challenge of learning and reasoning over language and vision data for the related downstream tasks such as visual question answering (VQA) and natural language for visual reasoning (NLVR).
no code implementations • 30 Apr 2020 • Fei Tang, Wanling Gao, Jianfeng Zhan, Chuanxin Lan, Xu Wen, Lei Wang, Chunjie Luo, Jiahui Dai, Zheng Cao, Xingwang Xiong, Zihan Jiang, Tianshu Hao, Fanda Fan, Fan Zhang, Yunyou Huang, Jianan Chen, Mengjia Du, Rui Ren, Chen Zheng, Daoyi Zheng, Haoning Tang, Kunlin Zhan, Biao Wang, Defei Kong, Minghe Yu, Chongkang Tan, Huan Li, Xinhui Tian, Yatao Li, Junchao Shao, Zhenyu Wang, Xiaoyu Wang, Hainan Ye
We use real-world benchmarks to cover the factors space that impacts the learning dynamics to the most considerable extent.
no code implementations • 17 Feb 2020 • Wanling Gao, Fei Tang, Jianfeng Zhan, Chuanxin Lan, Chunjie Luo, Lei Wang, Jiahui Dai, Zheng Cao, Xiongwang Xiong, Zihan Jiang, Tianshu Hao, Fanda Fan, Xu Wen, Fan Zhang, Yunyou Huang, Jianan Chen, Mengjia Du, Rui Ren, Chen Zheng, Daoyi Zheng, Haoning Tang, Kunlin Zhan, Biao Wang, Defei Kong, Minghe Yu, Chongkang Tan, Huan Li, Xinhui Tian, Yatao Li, Gang Lu, Junchao Shao, Zhenyu Wang, Xiaoyu Wang, Hainan Ye
An end-to-end benchmark is a distillation of the essential attributes of an industry-scale application.
no code implementations • 13 Aug 2019 • Wanling Gao, Fei Tang, Lei Wang, Jianfeng Zhan, Chunxin Lan, Chunjie Luo, Yunyou Huang, Chen Zheng, Jiahui Dai, Zheng Cao, Daoyi Zheng, Haoning Tang, Kunlin Zhan, Biao Wang, Defei Kong, Tong Wu, Minghe Yu, Chongkang Tan, Huan Li, Xinhui Tian, Yatao Li, Junchao Shao, Zhenyu Wang, Xiaoyu Wang, Hainan Ye
On the basis of the AIBench framework, abstracting the real-world data sets and workloads from one of the top e-commerce providers, we design and implement the first end-to-end Internet service AI benchmark, which contains the primary modules in the critical paths of an industry scale application and is scalable to deploy on different cluster scales.
no code implementations • 6 Aug 2019 • Tianshu Hao, Yunyou Huang, Xu Wen, Wanling Gao, Fan Zhang, Chen Zheng, Lei Wang, Hainan Ye, Kai Hwang, Zujie Ren, Jianfeng Zhan
In edge computing scenarios, the distribution of data and collaboration of workloads on different layers are serious concerns for performance, privacy, and security issues.
Performance Distributed, Parallel, and Cluster Computing
no code implementations • 22 Jun 2019 • Chen Zheng, Yu Sun, Shengxian Wan, dianhai yu
This paper proposes a novel End-to-End neural ranking framework called Reinforced Long Text Matching (RLTM) which matches a query with long documents efficiently and effectively.
3 code implementations • Chem. Mater. 2018 • Chi Chen, Weike Ye, Yunxing Zuo, Chen Zheng, Shyue Ping Ong
Similarly, we show that MEGNet models trained on $\sim 60, 000$ crystals in the Materials Project substantially outperform prior ML models in the prediction of the formation energies, band gaps and elastic moduli of crystals, achieving better than DFT accuracy over a much larger data set.
Ranked #4 on Formation Energy on Materials Project
Drug Discovery Formation Energy Materials Science Computational Physics
no code implementations • 23 Feb 2018 • Wanling Gao, Jianfeng Zhan, Lei Wang, Chunjie Luo, Daoyi Zheng, Xu Wen, Rui Ren, Chen Zheng, Xiwen He, Hainan Ye, Haoning Tang, Zheng Cao, Shujie Zhang, Jiahui Dai
On the basis of our previous work that identifies eight data motifs taking up most of the run time of a wide variety of big data and AI workloads, we propose a scalable benchmarking methodology that uses the combination of one or more data motifs---to represent diversity of big data and AI workloads.
no code implementations • 14 Nov 2017 • Chen Zheng, Shuangfei Zhai, Zhongfei Zhang
This approach uses the convolutional neural network and combines user feature representations with question feature representations to compute scores that the user who gets the highest score is the expert on this question.
no code implementations • 8 Nov 2017 • HuaChun Zhang, Lynden Kagan, Chen Zheng
In this paper, we proposed a new machine learning based fast power integrity classifier that quickly flags the EM/IR hotspots.
no code implementations • 6 Nov 2017 • Chen Zheng, Kiran Mathew, Chi Chen, Yiming Chen, Hanmei Tang, Alan Dozier, Joshua J. Kas, Fernando D. Vila, John J. Rehr, Louis F. J. Piper, Kristin Persson, Shyue Ping Ong
We report the development of XASdb, a large database of computed reference X-ray absorption spectra (XAS), and a novel Ensemble-Learned Spectra IdEntification (ELSIE) algorithm for the matching of spectra.
Materials Science
no code implementations • 28 Oct 2017 • Chen Zheng, Clara Grzegorz Kasprowicz, Carol Saunders
Customized routing optimizations are applied to the URNs and results show clear timing improvement and trend to converge toward timing closure.