Search Results for author: Zhengkai Tu

Found 8 papers, 6 papers with code

Beyond Major Product Prediction: Reproducing Reaction Mechanisms with Machine Learning Models Trained on a Large-Scale Mechanistic Dataset

no code implementations7 Mar 2024 Joonyoung F. Joung, Mun Hong Fong, Jihye Roh, Zhengkai Tu, John Bradshaw, Connor W. Coley

Mechanistic understanding of organic reactions can facilitate reaction development, impurity prediction, and in principle, reaction discovery.

Predictive Chemistry Augmented with Text Retrieval

1 code implementation8 Dec 2023 Yujie Qian, Zhening Li, Zhengkai Tu, Connor W. Coley, Regina Barzilay

Conventionally, chemoinformatics models are trained with extensive structured data manually extracted from the literature.

molecular representation Retrieval +2

RxnScribe: A Sequence Generation Model for Reaction Diagram Parsing

1 code implementation19 May 2023 Yujie Qian, Jiang Guo, Zhengkai Tu, Connor W. Coley, Regina Barzilay

Reaction diagram parsing is the task of extracting reaction schemes from a diagram in the chemistry literature.

Structured Prediction

FusionRetro: Molecule Representation Fusion via In-Context Learning for Retrosynthetic Planning

1 code implementation30 Sep 2022 Songtao Liu, Zhengkai Tu, Minkai Xu, Zuobai Zhang, Lu Lin, Rex Ying, Jian Tang, Peilin Zhao, Dinghao Wu

Current strategies use a decoupled approach of single-step retrosynthesis models and search algorithms, taking only the product as the input to predict the reactants for each planning step and ignoring valuable context information along the synthetic route.

Drug Discovery In-Context Learning +3

Permutation invariant graph-to-sequence model for template-free retrosynthesis and reaction prediction

1 code implementation19 Oct 2021 Zhengkai Tu, Connor W. Coley

Synthesis planning and reaction outcome prediction are two fundamental problems in computer-aided organic chemistry for which a variety of data-driven approaches have emerged.

Data Augmentation Graph-to-Sequence +5

Don't Change Me! User-Controllable Selective Paraphrase Generation

no code implementations EACL 2021 Mohan Zhang, Luchen Tan, Zhengkai Tu, Zihang Fu, Kun Xiong, Ming Li, Jimmy Lin

The contribution of this work is a novel data generation technique using distant supervision that allows us to start with a pretrained sequence-to-sequence model and fine-tune a paraphrase generator that exhibits this behavior, allowing user-controllable paraphrase generation.

Paraphrase Generation

Rapid Adaptation of BERT for Information Extraction on Domain-Specific Business Documents

1 code implementation5 Feb 2020 Ruixue Zhang, Wei Yang, Luyun Lin, Zhengkai Tu, Yuqing Xie, Zihang Fu, Yuhao Xie, Luchen Tan, Kun Xiong, Jimmy Lin

Techniques for automatically extracting important content elements from business documents such as contracts, statements, and filings have the potential to make business operations more efficient.

Cannot find the paper you are looking for? You can Submit a new open access paper.