no code implementations • 27 Jun 2024 • Hang Chen, Sheng Gao, Zejia Zhao, Zhengyang Duan, Haiou Zhang, Gordon Wetzstein, Xing Lin
Here, we propose an optical super-oscillatory diffractive neural network, i. e., SODNN, that can achieve super-resolved spatial resolution for imaging beyond the diffraction limit with superior performance over existing methods.
no code implementations • 19 Nov 2023 • Maria Cuellar, Sheng Gao, Heike Hofmann
To address this, we first generate a dataset of 3D toolmarks from various angles and directions using consecutively manufactured slotted screwdrivers.
no code implementations • 9 Jun 2023 • Hua Wang, Sheng Gao, Huanyu Zhang, Weijie J. Su, Milan Shen
In our paper, we introduce DP-HyPO, a pioneering framework for ``adaptive'' private hyperparameter optimization, aiming to bridge the gap between private and non-private hyperparameter optimization.
1 code implementation • 9 Dec 2022 • Ziyang Zheng, Zhengyang Duan, Hang Chen, Rui Yang, Sheng Gao, Haiou Zhang, Hongkai Xiong, Xing Lin
Photonic neural network (PNN) is a remarkable analog artificial intelligence (AI) accelerator that computes with photons instead of electrons to feature low latency, high energy efficiency, and high parallelism.
1 code implementation • 5 Aug 2022 • Jia Li, Ziyang Zhang, Junjie Lang, Yueqi Jiang, Liuwei An, Peng Zou, Yangyang Xu, Sheng Gao, Jie Lin, Chunxiao Fan, Xiao Sun, Meng Wang
In this paper, we present our solutions for the Multimodal Sentiment Analysis Challenge (MuSe) 2022, which includes MuSe-Humor, MuSe-Reaction and MuSe-Stress Sub-challenges.
1 code implementation • 9 Jun 2022 • Hua Wang, Sheng Gao, Huanyu Zhang, Milan Shen, Weijie J. Su
Many modern machine learning algorithms are composed of simple private algorithms; thus, an increasingly important problem is to efficiently compute the overall privacy loss under composition.
no code implementations • 1 Jul 2021 • Sheng Gao, Zongming Ma
Based on a Lagrangian form of the sample optimization problem, we propose a thresholded gradient descent algorithm for estimating GCA loading vectors and matrices in high dimensions.
1 code implementation • 28 Jun 2021 • Shanzhuo Zhang, Lihang Liu, Sheng Gao, Donglong He, Xiaomin Fang, Weibin Li, Zhengjie Huang, Weiyue Su, Wenjin Wang
In this report, we (SuperHelix team) present our solution to KDD Cup 2021-PCQM4M-LSC, a large-scale quantum chemistry dataset on predicting HOMO-LUMO gap of molecules.
1 code implementation • ACL 2021 • Jingxuan Yang, Kerui Xu, Jun Xu, Si Li, Sheng Gao, Jun Guo, Nianwen Xue, Ji-Rong Wen
A second (multi-relational) GCN is then applied to the utterance states to produce a discourse relation-augmented representation for the utterances, which are then fused together with token states in each utterance as input to a dropped pronoun recovery layer.
Ranked #5 on Discourse Parsing on STAC
1 code implementation • NAACL 2021 • Kun Liu, Yao Fu, Chuanqi Tan, Mosha Chen, Ningyu Zhang, Songfang Huang, Sheng Gao
This work studies NER under a noisy labeled setting with calibrated confidence estimation.
no code implementations • 1 Jan 2021 • Lulu Zhao, Zeyuan Yang, Weiran Xu, Sheng Gao, Jun Guo
In this paper, we present a Knowledge Graph Enhanced Dual-Copy network (KGEDC), a novel framework for abstractive dialogue summarization with conversational structure and factual knowledge.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jingxuan Yang, Kerui Xu, Jun Xu, Si Li, Sheng Gao, Jun Guo, Ji-Rong Wen, Nianwen Xue
Exploratory analysis also demonstrates that the GCRF did help to capture the dependencies between pronouns in neighboring utterances, thus contributes to the performance improvements.
no code implementations • IJCNLP 2019 • Guocheng Niu, Hengru Xu, Bolei He, Xinyan Xiao, Hua Wu, Sheng Gao
For text classification, traditional local feature driven models learn long dependency by deeply stacking or hybrid modeling.
1 code implementation • ACL 2019 • Kun Liu, Shen Li, Daqi Zheng, Zhengdong Lu, Sheng Gao, Si Li
To solve this problem, we propose a prism module to disentangle the semantic aspects of words and reduce noise at the input layer of a model.
Ranked #51 on Named Entity Recognition (NER) on CoNLL 2003 (English)
1 code implementation • NAACL 2019 • Jingxuan Yang, Jianzhuo Tong, Si Li, Sheng Gao, Jun Guo, Nianwen Xue
Pronouns are often dropped in Chinese sentences, and this happens more frequently in conversational genres as their referents can be easily understood from context.
no code implementations • 14 Apr 2019 • Jingxuan Yang, Jun Xu, Jianzhuo Tong, Sheng Gao, Jun Guo, Ji-Rong Wen
In the offline phase, IERT pre-trains deep item representations conditioning on their transaction contexts.
1 code implementation • CONLL 2018 • Hengru Xu, Shen Li, Renfen Hu, Si Li, Sheng Gao
Dropout is used to avoid overfitting by randomly dropping units from the neural networks during training.
no code implementations • NAACL 2018 • Chenliang Li, Weiran Xu, Si Li, Sheng Gao
Then, we introduce a Key Information Guide Network (KIGN), which encodes the keywords to the key information representation, to guide the process of generation.
Ranked #10 on Text Summarization on CNN / Daily Mail (Anonymized)
1 code implementation • 29 Apr 2018 • Yadi Lao, Jun Xu, Yanyan Lan, Jiafeng Guo, Sheng Gao, Xue-Qi Cheng
Inspired by the success and methodology of the AlphaGo Zero, MM-Tag formalizes the problem of sequence tagging with a Monte Carlo tree search (MCTS) enhanced Markov decision process (MDP) model, in which the time steps correspond to the positions of words in a sentence from left to right, and each action corresponds to assign a tag to a word.
no code implementations • WS 2017 • Zuyi Bao, Si Li, Weiran Xu, Sheng Gao
For Chinese word segmentation, the large-scale annotated corpora mainly focus on newswire and only a handful of annotated data is available in other domains such as patents and literature.
no code implementations • 24 Sep 2014 • Siting Ren, Sheng Gao
Cross-domain recommendation has been proposed to transfer user behavior pattern by pooling together the rating data from multiple domains to alleviate the sparsity problem appearing in single rating domains.