no code implementations • 3 Feb 2025 • YuHang Zhou, Giannis Karamanolakis, Victor Soto, Anna Rumshisky, Mayank Kulkarni, Furong Huang, Wei Ai, Jianhua Lu
The recent success of specialized Large Language Models (LLMs) in domains such as mathematical reasoning and coding has led to growing interest in methods for merging these expert LLMs into a unified Mixture-of-Experts (MoE) model, with the goal of enhancing performance in each domain while retaining effectiveness on general tasks.
no code implementations • 19 May 2024 • Sanchit Sinha, Yuguang Yue, Victor Soto, Mayank Kulkarni, Jianhua Lu, Aidong Zhang
In this paper, we propose MAML-en-LLM, a novel method for meta-training LLMs, which can learn truly generalizable parameters that not only perform well on disjointed tasks but also adapts to unseen tasks.
no code implementations • 14 Mar 2024 • Xiang Peng, Zhijin Qin, Xiaoming Tao, Jianhua Lu, Khaled B. Letaief
Semantic communications have gained significant attention as a promising approach to address the transmission bottleneck, especially with the continuous development of 6G techniques.
no code implementations • IEEE Transactions on Network Science and Engineering 2022 • Jiachen Sun, Ning Ge, Xu Chen, Wei Feng, Jianhua Lu
This screening algorithm is customer-oriented and offers personalized commodities by preventing unqualified sellers from participating in the transaction.
no code implementations • 8 Aug 2022 • Danlan Huang, Feifei Gao, Xiaoming Tao, Qiyuan Du, Jianhua Lu
Semantic communications has received growing interest since it can remarkably reduce the amount of data to be transmitted without missing critical information.
no code implementations • 15 Jun 2022 • Jack FitzGerald, Shankar Ananthakrishnan, Konstantine Arkoudas, Davide Bernardi, Abhishek Bhagia, Claudio Delli Bovi, Jin Cao, Rakesh Chada, Amit Chauhan, Luoxin Chen, Anurag Dwarakanath, Satyam Dwivedi, Turan Gojayev, Karthik Gopalakrishnan, Thomas Gueudre, Dilek Hakkani-Tur, Wael Hamza, Jonathan Hueser, Kevin Martin Jose, Haidar Khan, Beiye Liu, Jianhua Lu, Alessandro Manzotti, Pradeep Natarajan, Karolina Owczarzak, Gokmen Oz, Enrico Palumbo, Charith Peris, Chandana Satya Prakash, Stephen Rawls, Andy Rosenbaum, Anjali Shenoy, Saleh Soltan, Mukund Harakere Sridhar, Liz Tan, Fabian Triefenbach, Pan Wei, Haiyang Yu, Shuai Zheng, Gokhan Tur, Prem Natarajan
We present results from a large-scale experiment on pretraining encoders with non-embedding parameter counts ranging from 700M to 9. 3B, their subsequent distillation into smaller models ranging from 17M-170M parameters, and their application to the Natural Language Understanding (NLU) component of a virtual assistant system.
Cross-Lingual Natural Language Inference
intent-classification
+5
no code implementations • 6 Jun 2022 • Xiang Peng, Zhijin Qin, Danlan Huang, Xiaoming Tao, Jianhua Lu, Guangyi Liu, Chengkang Pan
With the advent of the 6G era, the concept of semantic communication has attracted increasing attention.
no code implementations • 30 Dec 2021 • Zhijin Qin, Xiaoming Tao, Jianhua Lu, Wen Tong, Geoffrey Ye Li
Semantic communication, regarded as the breakthrough beyond the Shannon paradigm, aims at the successful transmission of semantic information conveyed by the source rather than the accurate reception of each single symbol or bit regardless of its meaning.
no code implementations • 24 Nov 2021 • Khaled B. Letaief, Yuanming Shi, Jianmin Lu, Jianhua Lu
The thriving of artificial intelligence (AI) applications is driving the further evolution of wireless networks.
no code implementations • NAACL 2021 • Luoxin Chen, Francisco Garcia, Varun Kumar, He Xie, Jianhua Lu
This paper presents a production Semi-Supervised Learning (SSL) pipeline based on the student-teacher framework, which leverages millions of unlabeled examples to improve Natural Language Understanding (NLU) tasks.
no code implementations • 29 Mar 2021 • Zhiming Wang, Yantian Luo, Danlan Huang, Ning Ge, Jianhua Lu
Unsupervised domain adaptation (UDA) becomes more and more popular in tackling real-world problems without ground truth of the target domain.
no code implementations • 4 Mar 2021 • Chaoyi Han, Yiping Duan, Xiaoming Tao, Jianhua Lu
We show that our model performs well in measuring the similarity between restored and degraded images.
no code implementations • 21 Feb 2021 • Zhenyu Han, Fengli Xu, Yong Li, Tao Jiang, Depeng Jin, Jianhua Lu, James A. Evans
With the continued spread of coronavirus, the task of forecasting distinctive COVID-19 growth curves in different cities, which remain inadequately explained by standard epidemiological models, is critical for medical supply and treatment.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Luoxin Chen, Xinyue Liu, Weitong Ruan, Jianhua Lu
Adversarial training (AT) has shown strong regularization effects on deep learning algorithms by introducing small input perturbations to improve model robustness.
Ranked #3 on
Chunking
on CoNLL 2000
(using extra training data)
no code implementations • 17 Sep 2020 • Jianyu Cao, Wei Feng, Ning Ge, Jianhua Lu
Only a few research efforts have been devoted to other random delay characteristics, such as the delay bound violation probability and the probability distribution of the delay, by decoupling the transmission and computation processes of MEC.
no code implementations • ACL 2020 • Luoxin Chen, Weitong Ruan, Xinyue Liu, Jianhua Lu
Virtual adversarial training (VAT) is a powerful technique to improve model robustness in both supervised and semi-supervised settings.
Ranked #7 on
Chunking
on CoNLL 2000
no code implementations • 4 Jan 2016 • Xiangming Meng, Sheng Wu, Linling Kuang, Defeng, Huang, Jianhua Lu
We consider the problem of recovering clustered sparse signals with no prior knowledge of the sparsity pattern.