no code implementations • 6 Dec 2023 • Tingting Hou, Chang Jiang, Qing Lu
The advent of artificial intelligence, especially the progress of deep neural networks, is expected to revolutionize genetic research and offer unprecedented potential to decode the complex relationships between genetic variants and disease phenotypes, which could mark a significant step toward improving our understanding of the disease etiology.
no code implementations • 5 Dec 2023 • Tingting Hou, Chang Jiang, Qing Lu
While DNN plays a central role in modern AI technology, it has been rarely used in sequencing data analysis due to challenges brought by high-dimensional sequencing data (e. g., overfitting).
no code implementations • 30 Jul 2023 • Boyang Li, Bingyu Shen, Qing Lu, Taeho Jung, Yiyu Shi
In the conducted experiments, the PoFLSC consensus supported the subchain manager to be aware of reservation priority and the core partition of contributors to establish and maintain a competitive subchain.
no code implementations • 16 Dec 2022 • Xiaoxi Shen, Chang Jiang, Lyudmila Sakhanenko, Qing Lu
Neural networks (NN) play a central role in modern Artificial intelligence (AI) technology and has been successfully used in areas such as natural language processing and image recognition.
no code implementations • 20 Jun 2022 • Jinghang Lin, Shan Zhang, Qing Lu
Transfer learning has emerged as a powerful technique in many application problems, such as computer vision and natural language processing.
no code implementations • 8 Jun 2022 • Qing Lu, Xiaowei Xu, Shunjie Dong, Cong Hao, Lei Yang, Cheng Zhuo, Yiyu Shi
Accurately segmenting temporal frames of cine magnetic resonance imaging (MRI) is a crucial step in various real-time MRI guided cardiac interventions.
no code implementations • 5 May 2022 • Boyang Li, Qing Lu, Weiwen Jiang, Taeho Jung, Yiyu Shi
In many recent novel blockchain consensuses, the deep learning training procedure becomes the task for miners to prove their workload, thus the computation power of miners will not purely be spent on the hash puzzle.
no code implementations • 25 Apr 2021 • Wentao Chen, Hailong Qiu, Jian Zhuang, Chutong Zhang, Yu Hu, Qing Lu, Tianchen Wang, Yiyu Shi, Meiping Huang, Xiaowe Xu
Deep neural networks (DNNs) have demonstrated their great potential in recent years, exceeding the per-formance of human experts in a wide range of applications.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+5
no code implementations • 1 Jan 2021 • Qing Lu, Weiwen Jiang, Meng Jiang, Jingtong Hu, Sakyasingha Dasgupta, Yiyu Shi
The success of gragh neural networks (GNNs) in the past years has aroused grow-ing interest and effort in designing best models to handle graph-structured data.
no code implementations • 11 Nov 2020 • Muxuan Liang, Jaeyoung Park, Qing Lu, Xiang Zhong
The proposed method includes an MTL step using all outcomes to gain efficiency, and a subsequent calibration step using only the outcome of interest to correct both types of biases.
no code implementations • 26 Oct 2020 • Jinghang Lin, Xiaoran Tong, Chenxi Li, Qing Lu
Similar to expectile regression, ENN provides a comprehensive view of relationships between genetic variants and disease phenotypes and can be used to discover genetic variants predisposing to sub-populations (e. g., high-risk groups).
no code implementations • 30 Jan 2020 • Song Bian, Weiwen Jiang, Qing Lu, Yiyu Shi, Takashi Sato
Due to increasing privacy concerns, neural network (NN) based secure inference (SI) schemes that simultaneously hide the client inputs and server models attract major research interests.
no code implementations • 31 Oct 2019 • Qing Lu, Weiwen Jiang, Xiaowei Xu, Yiyu Shi, Jingtong Hu
With 30, 000 LUTs, a light-weight design is found to achieve 82. 98\% accuracy and 1293 images/second throughput, compared to which, under the same constraints, the traditional method even fails to find a valid solution.
no code implementations • 9 May 2019 • Yinglu Liu, Hao Shen, Yue Si, Xiaobo Wang, Xiangyu Zhu, Hailin Shi, Zhibin Hong, Hanqi Guo, Ziyuan Guo, Yanqin Chen, Bi Li, Teng Xi, Jun Yu, Haonian Xie, Guochen Xie, Mengyan Li, Qing Lu, Zengfu Wang, Shenqi Lai, Zhenhua Chai, Xiaoming Wei
However, previous competitions on facial landmark localization (i. e., the 300-W, 300-VW and Menpo challenges) aim to predict 68-point landmarks, which are incompetent to depict the structure of facial components.
no code implementations • CVPR 2018 • Xiaowei Xu, Qing Lu, Yu Hu, Lin Yang, Sharon Hu, Danny Chen, Yiyu Shi
Unlike existing litera- ture on quantization which primarily targets memory and computation complexity reduction, we apply quan- tization as a method to reduce over tting in FCNs for better accuracy.
1 code implementation • 4 Jan 2018 • Changshuai Wei, Qing Lu
Based on theoretical analysis, we proposed to use Laplacian kernel based similarity for GSU to boost power and enhance robustness.
no code implementations • 22 Jul 2017 • Qing Lu, Busheng Wang, Xiang-Rong Chen, Wu-Ming Liu
A large bulk band gap is critical for the applications of quantum spin hall (QSH) insulators in spintronics at room temperature.
Materials Science
no code implementations • 5 May 2015 • Changshuai Wei, Daniel J. Schaid, Qing Lu
Through simulation studies and an empirical data application, we found that TAMW outperformed multifactor dimensionality reduction (MDR) and the likelihood ratio based Mann whitney approach (LRMW) when the underlying complex disease involves multiple LME loci and their interactions.