no code implementations • 15 Jun 2023 • Lijun Yu, Jin Miao, Xiaoyu Sun, Jiayi Chen, Alexander G. Hauptmann, Hanjun Dai, Wei Wei
Visually-Rich Document Entity Retrieval (VDER) is a type of machine learning task that aims at recovering text spans in the documents for each of the entities in question.
no code implementations • 3 Jun 2023 • Kensen Shi, Hanjun Dai, Wen-Ding Li, Kevin Ellis, Charles Sutton
Search is an important technique in program synthesis that allows for adaptive strategies such as focusing on particular search directions based on execution results.
no code implementations • 26 May 2023 • Ruoxi Sun, Sercan O. Arik, Hootan Nakhost, Hanjun Dai, Rajarishi Sinha, Pengcheng Yin, Tomas Pfister
One impressive emergent capability of large language models (LLMs) is generation of code, including Structured Query Language (SQL) for databases.
no code implementations • 26 May 2023 • Dinghuai Zhang, Hanjun Dai, Nikolay Malkin, Aaron Courville, Yoshua Bengio, Ling Pan
Combinatorial optimization (CO) problems are often NP-hard and thus out of reach for exact algorithms, making them a tempting domain to apply machine learning methods.
no code implementations • 24 May 2023 • Xingchen Wan, Ruoxi Sun, Hootan Nakhost, Hanjun Dai, Julian Martin Eisenschlos, Sercan O. Arik, Tomas Pfister
A hallmark of modern large language models (LLMs) is their impressive general zero-shot and few-shot abilities, often elicited through prompt-based and/or in-context learning.
no code implementations • 23 May 2023 • Xingchen Wan, Ruoxi Sun, Hanjun Dai, Sercan O. Arik, Tomas Pfister
Modern large language models (LLMs) have demonstrated impressive capabilities at sophisticated tasks, often through step-by-step reasoning similar to humans.
no code implementations • 7 Mar 2023 • Azade Nova, Hanjun Dai, Dale Schuurmans
By only using the weights of the pre-trained model and unlabeled data, in a matter of a few minutes on a single GPU, up to 40% of the original FLOP count can be reduced with less than a 4% accuracy loss across all tasks considered.
no code implementations • 31 Jan 2023 • Yilun Du, Mengjiao Yang, Bo Dai, Hanjun Dai, Ofir Nachum, Joshua B. Tenenbaum, Dale Schuurmans, Pieter Abbeel
The proposed policy-as-video formulation can further represent environments with different state and action spaces in a unified space of images, which, for example, enables learning and generalization across a variety of robot manipulation tasks.
no code implementations • 30 Nov 2022 • Haoran Sun, Lijun Yu, Bo Dai, Dale Schuurmans, Hanjun Dai
Score-based modeling through stochastic differential equations (SDEs) has provided a new perspective on diffusion models, and demonstrated superior performance on continuous data.
no code implementations • 14 Nov 2022 • Hanjun Dai, Yuan Xue, Niao He, Bethany Wang, Na Li, Dale Schuurmans, Bo Dai
In real-world decision-making, uncertainty is important yet difficult to handle.
1 code implementation • 16 Sep 2022 • Haoran Sun, Hanjun Dai, Dale Schuurmans
Optimal scaling has been well studied for Metropolis-Hastings (M-H) algorithms in continuous spaces, but a similar understanding has been lacking in discrete spaces.
no code implementations • 23 Jul 2022 • Haoran Sun, Etash K. Guha, Hanjun Dai
However, learning neural networks for CO problems is notoriously difficult in lack of the labeled data as the training is easily trapped at local optima.
no code implementations • 13 Jul 2022 • Ruoxi Sun, Hanjun Dai, Adams Wei Yu
Extracting informative representations of molecules using Graph neural networks (GNNs) is crucial in AI-driven drug discovery.
no code implementations • 29 Jun 2022 • Haoran Sun, Hanjun Dai, Bo Dai, Haomin Zhou, Dale Schuurmans
It is known that gradient-based MCMC samplers for continuous spaces, such as Langevin Monte Carlo (LMC), can be derived as particle versions of a gradient flow that minimizes KL divergence on a Wasserstein manifold.
1 code implementation • ICLR 2022 • Kensen Shi, Hanjun Dai, Kevin Ellis, Charles Sutton
Many approaches to program synthesis perform a search within an enormous space of programs to find one that satisfies a given specification.
no code implementations • ICLR 2022 • Hanjun Dai, Yuan Xue, Zia Syed, Dale Schuurmans, Bo Dai
Stochastic dual dynamic programming (SDDP) is a state-of-the-art method for solving multi-stage stochastic optimization, widely used for modeling real-world process optimization tasks.
no code implementations • NeurIPS 2021 • Ruoxi Sun, Hanjun Dai, Li Li, Steven Kearnes, Bo Dai
In this paper, we propose a framework that unifies sequence- and graph-based methods as energy-based models (EBMs) with different energy functions.
1 code implementation • 28 Oct 2021 • Hongyu Ren, Hanjun Dai, Bo Dai, Xinyun Chen, Denny Zhou, Jure Leskovec, Dale Schuurmans
There are two important reasoning tasks on KGs: (1) single-hop knowledge graph completion, which involves predicting individual links in the KG; and (2), multi-hop reasoning, where the goal is to predict which KG entities satisfy a given logical query.
no code implementations • ICLR 2022 • Haoran Sun, Hanjun Dai, Wei Xia, Arun Ramamurthy
Energy-based Model (EBM) offers a powerful approach for modeling discrete structure, but both inference and learning of EBM are hard as it involves sampling from discrete distributions.
1 code implementation • ICLR 2022 • Pardis Pashakhanloo, Aaditya Naik, Yuepeng Wang, Hanjun Dai, Petros Maniatis, Mayur Naik
Designing a suitable representation for code-reasoning tasks is challenging in aspects such as the kinds of program information to model, how to combine them, and how much context to consider.
2 code implementations • NeurIPS 2021 • Hongyu Ren, Hanjun Dai, Zihang Dai, Mengjiao Yang, Jure Leskovec, Dale Schuurmans, Bo Dai
However, the key limitation of transformers is their quadratic memory and time complexity $\mathcal{O}(L^2)$ with respect to the sequence length in attention layers, which restricts application in extremely long sequences.
Ranked #2 on
Language Modelling
on Wiki-40B
1 code implementation • 26 Jun 2021 • Xinyun Chen, Petros Maniatis, Rishabh Singh, Charles Sutton, Hanjun Dai, Max Lin, Denny Zhou
In this work, we present the first approach for synthesizing spreadsheet formulas from tabular context, which includes both headers and semi-structured tabular data.
no code implementations • 1 Jan 2021 • Rongmei Lin, Hanjun Dai, Li Xiong, Wei Wei
We propose a generative fairness teaching framework that provides a model with not only real samples but also synthesized samples to compensate the data biases during training.
no code implementations • NeurIPS Workshop LMCA 2020 • Hanjun Dai, Xinshi Chen, Yu Li, Xin Gao, Le Song
Recently there is a surge of interests in using graph neural networks (GNNs) to learn algorithms.
no code implementations • 1 Jan 2021 • Binghong Chen, Chengtao Li, Hanjun Dai, Rampi Ramprasad, Le Song
We demonstrate that our method is able to propose high-quality polymerization plans for a dataset of 52 real-world polymers, of which a significant portion successfully recovers the currently-in-used polymerization processes in the real world.
no code implementations • ICLR 2021 • Binghong Chen, Tianzhe Wang, Chengtao Li, Hanjun Dai, Le Song
Optimizing molecules for desired properties is a fundamental yet challenging task in chemistry, material science and drug discovery.
no code implementations • NeurIPS 2020 • Yujia Xie, Hanjun Dai, Minshuo Chen, Bo Dai, Tuo Zhao, Hongyuan Zha, Wei Wei, Tomas Pfister
Finding the k largest or smallest elements from a collection of scores, i. e., top-k operation, is an important model component widely used in information retrieval, machine learning, and data mining.
no code implementations • NeurIPS 2020 • Hanjun Dai, Rishabh Singh, Bo Dai, Charles Sutton, Dale Schuurmans
In this paper we propose ALOE, a new algorithm for learning conditional and unconditional EBMs for discrete structured data, where parameter gradients are estimated using a learned sampler that mimics local search.
no code implementations • 4 Nov 2020 • Rohit Batra, Hanjun Dai, Tran Doan Huan, Lihua Chen, Chiho Kim, Will R. Gutekunst, Le Song, Rampi Ramprasad
The design/discovery of new materials is highly non-trivial owing to the near-infinite possibilities of material candidates, and multiple required property/performance objectives.
no code implementations • NeurIPS Workshop LMCA 2020 • Yujia Xie, Hanjun Dai, Minshuo Chen, Bo Dai, Tuo Zhao, Hongyuan Zha, Wei Wei, Tomas Pfister
The top-$k$ operation, i. e., finding the $k$ largest or smallest elements from a collection of scores, is an important model component, which is widely used in information retrieval, machine learning, and data mining.
no code implementations • ICLR 2021 • Augustus Odena, Kensen Shi, David Bieber, Rishabh Singh, Charles Sutton, Hanjun Dai
Program synthesis is challenging largely because of the difficulty of search in a large space of programs.
no code implementations • 14 Jul 2020 • Ruoxi Sun, Hanjun Dai, Li Li, Steven Kearnes, Bo Dai
Retrosynthesis -- the process of identifying a set of reactants to synthesize a target molecule -- is of vital importance to material design and drug discovery.
Ranked #1 on
Single-step retrosynthesis
on USPTO-50k
1 code implementation • ICML 2020 • Binghong Chen, Chengtao Li, Hanjun Dai, Le Song
Retrosynthetic planning is a critical task in organic chemistry which identifies a series of reactions that can lead to the synthesis of a target product.
Ranked #5 on
Multi-step retrosynthesis
on USPTO-190
1 code implementation • ICML 2020 • Hanjun Dai, Azade Nazi, Yujia Li, Bo Dai, Dale Schuurmans
Based on this, we develop a novel autoregressive model, named BiGG, that utilizes this sparsity to avoid generating the full adjacency matrix, and importantly reduces the graph generation time complexity to $O((n + m)\log n)$.
1 code implementation • ICML 2020 • Xinshi Chen, Hanjun Dai, Yu Li, Xin Gao, Le Song
Similar to algorithms, the optimal depth of a deep architecture may be different for different input instances, either to avoid ``over-thinking'', or because we want to compute less for operations converged already.
1 code implementation • ICLR 2020 • Elizabeth Dinella, Hanjun Dai, Ziyang Li, Mayur Naik, Le Song, Ke Wang
We present a learning-based approach to detect and fix a broad range of bugs in Javascript programs.
1 code implementation • ICML 2020 • Mengjiao Yang, Bo Dai, Hanjun Dai, Dale Schuurmans
Recently there has been growing interest in modeling sets with exchangeability such as point clouds.
no code implementations • 16 Feb 2020 • Yujia Xie, Hanjun Dai, Minshuo Chen, Bo Dai, Tuo Zhao, Hongyuan Zha, Wei Wei, Tomas Pfister
The top-k operation, i. e., finding the k largest or smallest elements from a collection of scores, is an important model component, which is widely used in information retrieval, machine learning, and data mining.
1 code implementation • NeurIPS 2019 • Hanjun Dai, Chengtao Li, Connor W. Coley, Bo Dai, Le Song
Retrosynthesis is one of the fundamental problems in organic chemistry.
Ranked #10 on
Single-step retrosynthesis
on USPTO-50k
no code implementations • NeurIPS 2019 • Hanjun Dai, Yujia Li, Chenglong Wang, Rishabh Singh, Po-Sen Huang, Pushmeet Kohli
We propose a `learning to explore' framework where we learn a policy from a distribution of environments.
no code implementations • NeurIPS 2018 • Harsh Shrivastava, Eugene Bart, Bob Price, Hanjun Dai, Bo Dai, Srinivas Aluru
We propose a new approach, called cooperative neural networks (CoNN), which uses a set of cooperatively trained neural networks to capture latent representations that exploit prior given independence structure.
no code implementations • ICLR 2019 • Xujie Si, Yuan Yang, Hanjun Dai, Mayur Naik, Le Song
Our framework consists of three components: 1) an encoder, which embeds both the logical specification and grammar at the same time using a graph neural network; 2) a grammar adaptive policy network which enables learning a transferable policy; and 3) a reinforcement learning algorithm that jointly trains the specification and grammar embedding and adaptive policy.
1 code implementation • NeurIPS 2019 • Bo Dai, Zhen Liu, Hanjun Dai, Niao He, Arthur Gretton, Le Song, Dale Schuurmans
We present an efficient algorithm for maximum likelihood estimation (MLE) of exponential family models, with a general parametrization of the energy function that includes neural networks.
2 code implementations • 2 Feb 2019 • Xinshi Chen, Hanjun Dai, Le Song
We present a particle flow realization of Bayes' rule, where an ODE-based neural operator is used to transport particles from a prior to its posterior after a new observation.
3 code implementations • 4 Dec 2018 • Thomas Kipf, Yujia Li, Hanjun Dai, Vinicius Zambaldi, Alvaro Sanchez-Gonzalez, Edward Grefenstette, Pushmeet Kohli, Peter Battaglia
We introduce Compositional Imitation Learning and Execution (CompILE): a framework for learning reusable, variable-length segments of hierarchically-structured behavior from demonstration data.
1 code implementation • NeurIPS 2018 • Xujie Si, Hanjun Dai, Mukund Raghothaman, Mayur Naik, Le Song
A fundamental problem in program verification concerns inferring loop invariants.
1 code implementation • NeurIPS 2018 • Bo Dai, Hanjun Dai, Niao He, Weiyang Liu, Zhen Liu, Jianshu Chen, Lin Xiao, Le Song
This flexible function class couples the variational distribution with the original parameters in the graphical models, allowing end-to-end learning of the graphical models by back-propagation through the variational distribution.
1 code implementation • 6 Nov 2018 • Bo Dai, Hanjun Dai, Arthur Gretton, Le Song, Dale Schuurmans, Niao He
We investigate penalized maximum log-likelihood estimation for exponential family distributions whose natural parameter resides in a reproducing kernel Hilbert space.
no code implementations • ICML 2018 • Hanjun Dai, Zornitsa Kozareva, Bo Dai, Alex Smola, Le Song
Many graph analytics problems can be solved via iterative algorithms where the solutions are often characterized by a set of steady-state conditions.
1 code implementation • ICML 2018 • Hanjun Dai, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, Le Song
Deep learning on graph structures has shown exciting results in various applications.
no code implementations • 31 May 2018 • Yuyu Zhang, Hanjun Dai, Kamil Toraman, Le Song
Our model learns to reason with neural embeddings of both knowledge graphs.
1 code implementation • ICLR 2018 • Hanjun Dai, Yingtao Tian, Bo Dai, Steven Skiena, Le Song
Deep generative models have been enjoying success in modeling continuous data.
1 code implementation • 12 Sep 2017 • Yuyu Zhang, Hanjun Dai, Zornitsa Kozareva, Alexander J. Smola, Le Song
Knowledge graph (KG) is known to be helpful for the task of question answering (QA), since it provides well-structured relational information between entities, and allows one to further infer indirect facts.
2 code implementations • ICML 2017 • Rakshit Trivedi, Hanjun Dai, Yichen Wang, Le Song
The occurrence of a fact (edge) is modeled as a multivariate point process whose intensity function is modulated by the score for that fact computed based on the learned entity embeddings.
8 code implementations • NeurIPS 2017 • Hanjun Dai, Elias B. Khalil, Yuyu Zhang, Bistra Dilkina, Le Song
The design of good heuristics or approximation algorithms for NP-hard combinatorial optimization problems often requires significant specialized knowledge and trial-and-error.
no code implementations • 13 Sep 2016 • Hanjun Dai, Yichen Wang, Rakshit Trivedi, Le Song
DeepCoevolve use recurrent neural network (RNN) over evolving networks to define the intensity function in point processes, which allows the model to capture complex mutual influence between users and items, and the feature evolution over time.
1 code implementation • 17 Mar 2016 • Hanjun Dai, Bo Dai, Le Song
Kernel classifiers and regressors designed for structured data, such as sequences, trees and graphs, have significantly advanced a number of interdisciplinary areas such as computational biology and drug design.
no code implementations • NeurIPS 2015 • Shuang Li, Yao Xie, Hanjun Dai, Le Song
Detecting the emergence of an abrupt change-point is a classic problem in statistics and machine learning.
no code implementations • 1 Sep 2015 • Yao Xie, Ruiyang Song, Hanjun Dai, Qingbin Li, Le Song
The optimization problem for OSDR is non-convex and hard to analyze in general; we provide convergence analysis of OSDR in a simple linear regression setting.
no code implementations • 5 Jul 2015 • Shuang Li, Yao Xie, Hanjun Dai, Le Song
A novel theoretical result of the paper is the characterization of the tail probability of these statistics using the change-of-measure technique, which focuses on characterizing the tail of the detection statistics rather than obtaining its asymptotic distribution under the null distribution.
no code implementations • 9 Jun 2015 • Bo Dai, Niao He, Hanjun Dai, Le Song
Bayesian methods are appealing in their flexibility in modeling complex data and ability in capturing uncertainty in parameters.
no code implementations • 23 Apr 2014 • Yuyu Zhang, Hanjun Dai, Chang Xu, Jun Feng, Taifeng Wang, Jiang Bian, Bin Wang, Tie-Yan Liu
Click prediction is one of the fundamental problems in sponsored search.