no code implementations • CONSTRAINT (ACL) 2022 • Ziming Zhou, Han Zhao, Jingjing Dong, Jun Gao, Xiaolong Liu
The memes serve as an important tool in online communication, whereas some hateful memes endanger cyberspace by attacking certain people or subjects.
no code implementations • Findings (EMNLP) 2021 • Jun Gao, YuHan Liu, Haolin Deng, Wei Wang, Yu Cao, Jiachen Du, Ruifeng Xu
The emotion cause is a stimulus for human emotions.
1 code implementation • 5 Mar 2025 • Xuanchi Ren, Tianchang Shen, Jiahui Huang, Huan Ling, Yifan Lu, Merlin Nimier-David, Thomas Müller, Alexander Keller, Sanja Fidler, Jun Gao
Our results demonstrate more precise camera control than prior work, as well as state-of-the-art results in sparse-view novel view synthesis, even in challenging settings such as driving scenes and monocular dynamic video.
no code implementations • 3 Mar 2025 • Jay Zhangjie Wu, Yuxuan Zhang, Haithem Turki, Xuanchi Ren, Jun Gao, Mike Zheng Shou, Sanja Fidler, Zan Gojcic, Huan Ling
At the core of our approach is Difix, a single-step image diffusion model trained to enhance and remove artifacts in rendered novel views caused by underconstrained regions of the 3D representation.
no code implementations • 30 Jan 2025 • Ruofan Liang, Zan Gojcic, Huan Ling, Jacob Munkberg, Jon Hasselgren, Zhi-Hao Lin, Jun Gao, Alexander Keller, Nandita Vijaykumar, Sanja Fidler, Zian Wang
Classic physically-based rendering (PBR) accurately simulates the light transport, but relies on precise scene representations--explicit 3D geometry, high-quality material properties, and lighting conditions--that are often impractical to obtain in real-world scenarios.
no code implementations • 5 Dec 2024 • Yifan Lu, Xuanchi Ren, Jiawei Yang, Tianchang Shen, Zhangjie Wu, Jun Gao, Yue Wang, Siheng Chen, Mike Chen, Sanja Fidler, Jiahui Huang
We present InfiniCube, a scalable method for generating unbounded dynamic 3D driving scenes with high fidelity and controllability.
no code implementations • 29 Nov 2024 • Jun Gao, Yongqi Li, Ziqiang Cao, Wenjie Li
Chain-of-Thought (CoT) prompting elicits large language models (LLMs) to produce a series of intermediate reasoning steps before arriving at the final answer.
1 code implementation • 30 Oct 2024 • Shengkai Zhang, Nianhong Jiao, Tian Li, Chaojie Yang, Chenhui Xue, Boya Niu, Jun Gao
We propose an effective method for inserting adapters into text-to-image foundation models, which enables the execution of complex downstream tasks while preserving the generalization ability of the base model.
no code implementations • 7 Oct 2024 • Hao Miao, Zida Liu, Jun Gao
Motivated by these limitations, this paper proposes a method named BSG4Bot with an intuition that GNNs training on Biased SubGraphs can improve both performance and time/space efficiency in bot detection.
no code implementations • 30 Sep 2024 • Tianchang Shen, Zhaoshuo Li, Marc Law, Matan Atzmon, Sanja Fidler, James Lucas, Jun Gao, Nicholas Sharp
In particular, our vertex embeddings generate cyclic neighbor relationships in a halfedge mesh representation, which gives a guarantee of edge-manifoldness and the ability to represent general polygonal meshes.
no code implementations • 13 Sep 2024 • Yikuan Xia, Jiazun Chen, Jun Gao
The challenge is to build a RAG system from web sources and knowledge graphs.
1 code implementation • 1 Aug 2024 • Yu Xie, Qian Qiao, Jun Gao, Tianxiang Wu, Jiaqing Fan, Yue Zhang, Jielei Zhang, Huyang Sun
Unfortunately, this denoising training method cannot be directly applied to text spotting tasks, as these tasks need to perform irregular shape detection tasks and more complex text recognition tasks than classification.
no code implementations • 11 Jun 2024 • Jun Gao, Qian Qiao, Ziqiang Cao, Zili Wang, Wenjie Li
In-context learning (ICL) facilitates Large Language Models (LLMs) exhibiting emergent ability on downstream tasks without updating billions of parameters.
no code implementations • 3 Jun 2024 • Jun Gao, Ziqiang Cao, Shaoyao Huang, Luozheng Qin, Chunhui Ai
Then, we require ChatGPT to generate $k$ candidate summaries for the inference document at a time under the guidance of the retrieved demonstration.
no code implementations • 27 May 2024 • Jun Gao, Ziqiang Cao, Wenjie Li
In this paper, we propose an ICL framework UniICL, which Unifies demonstration selection and compression, and final response generation via a single frozen LLM.
no code implementations • 27 May 2024 • Jun Gao, Ziqiang Cao, Wenjie Li
This paper proposes a Self-Compressor (SelfCP), which employs the target LLM itself to compress over-limit prompts into dense vectors while keeping the allowed prompts unmodified.
no code implementations • 8 May 2024 • Yikuan Xia, Jiazun Chen, Xinchi Li, Jun Gao
The first is an augmented contextualized soft token-based prompt tuning method that extracts a guiding soft token benefit for the PLMs' prompt tuning, and the second is a cost-effective information augmentation strategy leveraging large language models (LLMs).
no code implementations • 22 Mar 2024 • Kevin Xie, Jonathan Lorraine, Tianshi Cao, Jun Gao, James Lucas, Antonio Torralba, Sanja Fidler, Xiaohui Zeng
Recent text-to-3D generation approaches produce impressive 3D results but require time-consuming optimization that can take up to an hour per prompt.
no code implementations • 4 Mar 2024 • Chen Xu, Tian Lan, Changlong Yu, Wei Wang, Jun Gao, Yu Ji, Qunxi Dong, Kun Qian, Piji Li, Wei Bi, Bin Hu
Constrained decoding approaches aim to control the meaning or style of text generated by a Pre-trained Language Model (PLM) using specific target words during inference.
1 code implementation • 4 Mar 2024 • Haolin Deng, Chang Wang, Xin Li, Dezhang Yuan, Junlang Zhan, Tianhua Zhou, Jin Ma, Jun Gao, Ruifeng Xu
Enhancing the attribution in large language models (LLMs) is a crucial task.
1 code implementation • 24 Feb 2024 • Zekun Jiang, Dongjie Cheng, Ziyuan Qin, Jun Gao, Qicheng Lao, Abdullaev Bakhrom Ismoilovich, Urazboev Gayrat, Yuldashov Elyorbek, Bekchanov Habibullo, Defu Tang, LinJing Wei, Kang Li, Le Zhang
This study presents a novel multimodal medical image zero-shot segmentation algorithm named the text-visual-prompt segment anything model (TV-SAM) without any manual annotations.
1 code implementation • 18 Feb 2024 • Jun Gao, Huan Zhao, Wei Wang, Changlong Yu, Ruifeng Xu
In this study, we present EventRL, a reinforcement learning approach developed to enhance event extraction for large language models (LLMs).
no code implementations • 18 Feb 2024 • Lanning Wei, Jun Gao, Huan Zhao, Quanming Yao
This paper proposes a novel conceptual prototype for designing versatile graph learning methods with LLMs, with a particular focus on the "where" and "how" perspectives.
no code implementations • 22 Nov 2023 • Katja Schwarz, Seung Wook Kim, Jun Gao, Sanja Fidler, Andreas Geiger, Karsten Kreis
Then, we train a diffusion model in the 3D-aware latent space, thereby enabling synthesis of high-quality 3D-consistent image samples, outperforming recent state-of-the-art GAN-based methods.
no code implementations • 17 Nov 2023 • Zhaoyang Ma, Xupeng He, Shuyu Sun, BiCheng Yan, Hyung Kwak, Jun Gao
Despite its advanced features, SAM struggles with rock CT/SEM images due to their absence in its training set and the low-contrast nature of grayscale images.
no code implementations • 16 Nov 2023 • Zian Wang, Tianchang Shen, Merlin Nimier-David, Nicholas Sharp, Jun Gao, Alexander Keller, Sanja Fidler, Thomas Müller, Zan Gojcic
We then extract an explicit mesh of a narrow band around the surface, with width determined by the kernel size, and fine-tune the radiance field within this band.
no code implementations • 10 Nov 2023 • Zhaoyang Ma, Xupeng He, Hyung Kwak, Jun Gao, Shuyu Sun, BiCheng Yan
In digital rock physics, analysing microstructures from CT and SEM scans is crucial for estimating properties like porosity and pore connectivity.
no code implementations • 8 Oct 2023 • Jun Gao, Huan Zhao, Yice Zhang, Wei Wang, Changlong Yu, Ruifeng Xu
Information Extraction (IE) is an essential task in Natural Language Processing.
no code implementations • 30 Sep 2023 • Chengdong Ma, Ziran Yang, Hai Ci, Jun Gao, Minquan Gao, Xuehai Pan, Yaodong Yang
Furthermore, we develop a Gamified Red Team Solver (GRTS) with diversity measures to mitigate mode collapse and theoretically guarantee the convergence of approximate Nash equilibrium which results in better strategies for both teams.
1 code implementation • 30 Sep 2023 • Zhaowei Zhang, Fengshuo Bai, Jun Gao, Yaodong Yang
Personal values are a crucial factor behind human decision-making.
no code implementations • 28 Aug 2023 • Dongze Wu, Jun Gao, Feng Yin
Existing localization approaches utilizing environment-specific channel state information (CSI) excel under specific environment but struggle to generalize across varied environments.
1 code implementation • 10 Aug 2023 • Tianchang Shen, Jacob Munkberg, Jon Hasselgren, Kangxue Yin, Zian Wang, Wenzheng Chen, Zan Gojcic, Sanja Fidler, Nicholas Sharp, Jun Gao
This work considers gradient-based mesh optimization, where we iteratively optimize for a 3D surface mesh by representing it as the isosurface of a scalar field, an increasingly common paradigm in applications including photogrammetry, generative modeling, and inverse physics.
1 code implementation • 14 Jun 2023 • Qingbo Kang, Jun Gao, Kang Li, Qicheng Lao
Overall, our work highlights the potential of MAE for ultrasound image recognition and presents a novel approach that incorporates deblurring to further improve its effectiveness.
1 code implementation • 27 May 2023 • Fangqi Zhu, Lin Zhang, Jun Gao, Bing Qin, Ruifeng Xu, Haiqin Yang
Event skeleton generation, aiming to induce an event schema skeleton graph with abstracted event nodes and their temporal relations from a set of event instance graphs, is a critical step in the temporal complex event schema induction task.
no code implementations • 18 May 2023 • Aysegul Dundar, Jun Gao, Andrew Tao, Bryan Catanzaro
In this work, to overcome these limitations of generated datasets, we have two main contributions which lead us to achieve state-of-the-art results on challenging objects: 1) A robust multi-stage learning scheme that gradually relies more on the models own predictions when calculating losses, 2) A novel adversarial learning pipeline with online pseudo-ground truth generations to achieve fine details.
no code implementations • 6 Apr 2023 • Zian Wang, Tianchang Shen, Jun Gao, Shengyu Huang, Jacob Munkberg, Jon Hasselgren, Zan Gojcic, Wenzheng Chen, Sanja Fidler
Reconstruction and intrinsic decomposition of scenes from captured imagery would enable many applications such as relighting and virtual object insertion.
no code implementations • 7 Mar 2023 • Jun Gao, Huan Zhao, Changlong Yu, Ruifeng Xu
While ChatGPT has demonstrated impressive results in tasks like machine translation, text summarization, and question answering, it presents challenges when used for complex tasks like event extraction.
no code implementations • 6 Jan 2023 • Jun Gao, Changlong Yu, Wei Wang, Huan Zhao, Ruifeng Xu
We present Mask-then-Fill, a flexible and effective data augmentation framework for event extraction.
no code implementations • CVPR 2023 • Zian Wang, Tianchang Shen, Jun Gao, Shengyu Huang, Jacob Munkberg, Jon Hasselgren, Zan Gojcic, Wenzheng Chen, Sanja Fidler
Reconstruction and intrinsic decomposition of scenes from captured imagery would enable many applications such as relighting and virtual object insertion.
1 code implementation • 7 Dec 2022 • Fangqi Zhu, Jun Gao, Changlong Yu, Wei Wang, Chen Xu, Xin Mu, Min Yang, Ruifeng Xu
First, the pretrained language models adopted by current works ignore event-level knowledge, resulting in an inability to capture the correlations between events well.
1 code implementation • CVPR 2023 • Chen-Hsuan Lin, Jun Gao, Luming Tang, Towaki Takikawa, Xiaohui Zeng, Xun Huang, Karsten Kreis, Sanja Fidler, Ming-Yu Liu, Tsung-Yi Lin
DreamFusion has recently demonstrated the utility of a pre-trained text-to-image diffusion model to optimize Neural Radiance Fields (NeRF), achieving remarkable text-to-3D synthesis results.
Ranked #2 on
Text to 3D
on T$^3$Bench
no code implementations • 8 Nov 2022 • Jun Gao, Dongze Wu, Feng Yin, Qinglei Kong, Lexi Xu, Shuguang Cui
The framework introduces two paradigms for the optimization of meta-parameters: a centralized paradigm that simplifies the process by sharing data from all historical environments, and a distributed paradigm that maintains data privacy by training meta-parameters for each specific environment separately.
no code implementations • 2 Nov 2022 • Haolin Deng, Yanan Zhang, Yangfan Zhang, Wangyang Ying, Changlong Yu, Jun Gao, Wei Wang, Xiaoling Bai, Nan Yang, Jin Ma, Xiang Chen, Tianhua Zhou
To the best of our knowledge, it is currently the largest manually-annotated Chinese dataset for open event extraction.
1 code implementation • 27 Oct 2022 • Yu Cao, Dianqi Li, Meng Fang, Tianyi Zhou, Jun Gao, Yibing Zhan, DaCheng Tao
We present Twin Answer Sentences Attack (TASA), an adversarial attack method for question answering (QA) models that produces fluent and grammatical adversarial contexts while maintaining gold answers.
3 code implementations • 22 Sep 2022 • Jun Gao, Tianchang Shen, Zian Wang, Wenzheng Chen, Kangxue Yin, Daiqing Li, Or Litany, Zan Gojcic, Sanja Fidler
As several industries are moving towards modeling massive 3D virtual worlds, the need for content creation tools that can scale in terms of the quantity, quality, and diversity of 3D content is becoming evident.
no code implementations • 5 Jul 2022 • Gary Leung, Jun Gao, Xiaohui Zeng, Sanja Fidler
HILA extends hierarchical vision transformer architectures by adding local connections between features of higher and lower levels to the backbone encoder.
1 code implementation • 9 Jun 2022 • Wentao Zhang, Zeang Sheng, Ziqi Yin, Yuezihan Jiang, Yikuan Xia, Jun Gao, Zhi Yang, Bin Cui
Graph Neural Networks (GNNs) have achieved great success in various graph mining tasks. However, drastic performance degradation is always observed when a GNN is stacked with many layers.
1 code implementation • NAACL 2022 • Hanhao Qu, Yu Cao, Jun Gao, Liang Ding, Ruifeng Xu
We present IBR, an Iterative Backward Reasoning model to solve the proof generation tasks on rule-based Question Answering (QA), where models are required to reason over a series of textual rules and facts to find out the related proof path and derive the final answer.
1 code implementation • 7 May 2022 • YuHan Liu, Jun Gao, Jiachen Du, Lanjun Zhou, Ruifeng Xu
The emotion-aware dialogue management contains two parts: (1) Emotion state tracking maintains the current emotion state of the user and (2) Empathetic dialogue policy selection predicts a target emotion and a user's intent based on the results of the emotion state tracking.
no code implementations • 29 Mar 2022 • Kunyuan Li, Jun Zhang, Jun Gao, Meibin Qi
In this paper, we propose a self-supervised learning framework for light field depth estimation.
no code implementations • 17 Mar 2022 • Aysegul Dundar, Jun Gao, Andrew Tao, Bryan Catanzaro
The reconstruction is posed as an adaptation problem and is done progressively where in the first stage, we focus on learning accurate geometry, whereas in the second stage, we focus on learning the texture with a generative adversarial network.
1 code implementation • ACL 2022 • Jun Gao, Wei Wang, Changlong Yu, Huan Zhao, Wilfred Ng, Ruifeng Xu
Representations of events described in text are important for various tasks.
2 code implementations • CVPR 2022 • Jacob Munkberg, Jon Hasselgren, Tianchang Shen, Jun Gao, Wenzheng Chen, Alex Evans, Thomas Müller, Sanja Fidler
We present an efficient method for joint optimization of topology, materials and lighting from multi-view image observations.
Ranked #2 on
Depth Prediction
on Stanford-ORB
no code implementations • NeurIPS 2021 • Tianchang Shen, Jun Gao, Kangxue Yin, Ming-Yu Liu, Sanja Fidler
The core of DMTet includes a deformable tetrahedral grid that encodes a discretized signed distance function and a differentiable marching tetrahedra layer that converts the implicit signed distance representation to the explicit surface mesh representation.
no code implementations • NeurIPS 2021 • Wenzheng Chen, Joey Litalien, Jun Gao, Zian Wang, Clement Fuji Tsang, Sameh Khamis, Or Litany, Sanja Fidler
We consider the challenging problem of predicting intrinsic object properties from a single image by exploiting differentiable renderers.
no code implementations • ICCV 2021 • Kangxue Yin, Jun Gao, Maria Shugrina, Sameh Khamis, Sanja Fidler
Given a small set of high-quality textured objects, our method can create many novel stylized shapes, resulting in effortless 3D content creation and style-ware data augmentation.
1 code implementation • 2 Aug 2021 • Wentao Zhang, Zeang Sheng, Yuezihan Jiang, Yikuan Xia, Jun Gao, Zhi Yang, Bin Cui
Based on the experimental results, we answer the following two essential questions: (1) what actually leads to the compromised performance of deep GNNs; (2) when we need and how to build deep GNNs.
no code implementations • 30 May 2021 • Jun Gao, Wei Bi, Ruifeng Xu, Shuming Shi
We first clarify an assumption on reference-based metrics that, if more high-quality references are added into the reference set, the reliability of the metric will increase.
2 code implementations • CVPR 2021 • Yuxuan Zhang, Huan Ling, Jun Gao, Kangxue Yin, Jean-Francois Lafleche, Adela Barriuso, Antonio Torralba, Sanja Fidler
To showcase the power of our approach, we generated datasets for 7 image segmentation tasks which include pixel-level labels for 34 human face parts, and 32 car parts.
1 code implementation • NeurIPS 2020 • Jun Gao, Wenzheng Chen, Tommy Xiang, Clement Fuji Tsang, Alec Jacobson, Morgan McGuire, Sanja Fidler
We introduce Deformable Tetrahedral Meshes (DefTet) as a particular parameterization that utilizes volumetric tetrahedral meshes for the reconstruction problem.
no code implementations • ICLR 2021 • Yuxuan Zhang, Wenzheng Chen, Huan Ling, Jun Gao, Yinan Zhang, Antonio Torralba, Sanja Fidler
Key to our approach is to exploit GANs as a multi-view data generator to train an inverse graphics network using an off-the-shelf differentiable renderer, and the trained inverse graphics network as a teacher to disentangle the GAN's latent code into interpretable 3D properties.
no code implementations • ICLR 2021 • Xinran Wang, Yu Xiang, Jun Gao, Jie Ding
In this work, we propose information laundering, a novel framework for enhancing model privacy.
no code implementations • ECCV 2020 • Tianchang Shen, Jun Gao, Amlan Kar, Sanja Fidler
We implement our framework as a web service and conduct a user study, where we show that user annotated data using our method effectively facilitates real-world learning tasks.
no code implementations • ECCV 2020 • Jun Gao, Zian Wang, Jinchen Xuan, Sanja Fidler
We also utilize DefGrid at the output layers for the task of object mask annotation, and show that reasoning about object boundaries on our predicted polygonal grid leads to more accurate results over existing pixel-wise and curve-based approaches.
no code implementations • 6 Aug 2020 • Hao Tang, Anurag Pal, Lu-Feng Qiao, Tian-Yu Wang, Jun Gao, Xian-Min Jin
Collateralized debt obligation (CDO) has been one of the most commonly used structured financial products and is intensively studied in quantitative finance.
1 code implementation • 9 Jul 2020 • Kunyuan Li, Jun Zhang, Rui Sun, Xu-Dong Zhang, Jun Gao
Based on the observation that an oriented line and its neighboring pixels in an EPI share a similar linear structure, we propose an end-to-end fully convolutional network (FCN) to estimate the depth value of the intersection point on the horizontal and vertical EPIs.
no code implementations • IJCNLP 2019 • Jun Gao, Wei Bi, Xiaojiang Liu, Junhui Li, Guodong Zhou, Shuming Shi
In this paper, we introduce a discrete latent variable with an explicit semantic meaning to improve the CVAE on short-text conversation.
no code implementations • 6 Aug 2019 • Jun Gao, Luyun Gan, Fabiola Buschendorf, Liao Zhang, Hua Liu, Peixue Li, Xiaodai Dong, Tao Lu
We investigate deep learning based omni intrusion detection system (IDS) for supervisory control and data acquisition (SCADA) networks that are capable of detecting both temporally uncorrelated and correlated attacks.
1 code implementation • NeurIPS 2019 • Wenzheng Chen, Jun Gao, Huan Ling, Edward J. Smith, Jaakko Lehtinen, Alec Jacobson, Sanja Fidler
Many machine learning models operate on images, but ignore the fact that images are 2D projections formed by 3D geometry interacting with light, in a process called rendering.
Ranked #4 on
Single-View 3D Reconstruction
on ShapeNet
1 code implementation • ICLR 2019 • Jun Gao, Di He, Xu Tan, Tao Qin, Li-Wei Wang, Tie-Yan Liu
We study an interesting problem in training neural network-based models for natural language generation tasks, which we call the \emph{representation degeneration problem}.
no code implementations • ACL 2019 • Wei Bi, Jun Gao, Xiaojiang Liu, Shuming Shi
Classification models are trained on this dataset to (i) recognize the sentence function of new data in a large corpus of short-text conversations; (ii) estimate a proper sentence function of the response given a test query.
no code implementations • 3 Apr 2019 • Jinze Bai, Chang Zhou, Junshuai Song, Xiaoru Qu, Weiting An, Zhao Li, Jun Gao
In particular, BGN improves the precision of the best competitors by 16\% on average while maintaining the highest diversity on four datasets, and yields a 3. 85x improvement of response time over the best competitors in the bundle list recommendation problem.
no code implementations • 27 Mar 2019 • Jun Gao, Xiao Li, Li-Wei Wang, Sanja Fidler, Stephen Lin
We present a method for compositing virtual objects into a photograph such that the object colors appear to have been processed by the photo's camera imaging pipeline.
2 code implementations • CVPR 2019 • Huan Ling, Jun Gao, Amlan Kar, Wenzheng Chen, Sanja Fidler
Our model runs at 29. 3ms in automatic, and 2. 6ms in interactive mode, making it 10x and 100x faster than Polygon-RNN++.
2 code implementations • 12 Jan 2019 • Jun Gao, Chengcheng Tang, Vignesh Ganapathi-Subramanian, Jiahui Huang, Hao Su, Leonidas J. Guibas
Reconstruction of geometry based on different input modes, such as images or point clouds, has been instrumental in the development of computer aided design and computer graphics.
no code implementations • 14 Nov 2018 • Jun Gao, Wei Bi, Xiaojiang Liu, Junhui Li, Shuming Shi
In this paper, we propose a novel response generation model, which considers a set of responses jointly and generates multiple diverse responses simultaneously.
1 code implementation • EMNLP 2018 • Yahui Liu, Wei Bi, Jun Gao, Xiaojiang Liu, Jian Yao, Shuming Shi
We observe that in the conversation tasks, each query could have multiple responses, which forms a 1-to-n or m-to-n relationship in the view of the total corpus.
12 code implementations • ECCV 2018 • Ze Yang, Tiange Luo, Dong Wang, Zhiqiang Hu, Jun Gao, Li-Wei Wang
In consideration of intrinsic consistency between informativeness of the regions and their probability being ground-truth class, we design a novel training paradigm, which enables Navigator to detect most informative regions under the guidance from Teacher.
Ranked #45 on
Fine-Grained Image Classification
on FGVC Aircraft
no code implementations • 26 Aug 2018 • Ci-Yu Wang, Jun Gao, Zhi-Qiang Jiao, Lu-Feng Qiao, Ruo-Jing Ren, Zhen Feng, Yuan Chen, Zeng-Quan Yan, Yao Wang, Hao Tang, Xian-Min Jin
Quantum key distribution (QKD), harnessing quantum physics and optoelectronics, may promise unconditionally secure information exchange in theory.
Quantum Physics
no code implementations • ICML 2018 • Wenlong Mou, Yuchen Zhou, Jun Gao, Li-Wei Wang
We study the problem of generalization guarantees for dropout training.
2 code implementations • 17 Nov 2017 • Chang Zhou, Jinze Bai, Junshuai Song, Xiaofei Liu, Zhengchao Zhao, Xiusi Chen, Jun Gao
Downstream applications then can use the user behavior vectors via vanilla attention.
no code implementations • 28 May 2017 • Fang Liu, Likai Du, Dongju Zhang, Jun Gao
Based on the knowledge of these meta-stable patterns, we suggested an interpolation scheme with only a concrete and finite set of known patterns to accurately predict the ground and excited state properties of the entire dynamics trajectories.