Search Results for author: Mingchen Li

Found 32 papers, 12 papers with code

PETA: Evaluating the Impact of Protein Transfer Learning with Sub-word Tokenization on Downstream Applications

1 code implementation26 Oct 2023 Yang Tan, Mingchen Li, Pan Tan, Ziyi Zhou, Huiqun Yu, Guisheng Fan, Liang Hong

Moreover, despite the wealth of benchmarks and studies in the natural language community, there remains a lack of a comprehensive benchmark for systematically evaluating protein language model quality.

Protein Language Model Transfer Learning

DynaMo: Dynamic Community Detection by Incrementally Maximizing Modularity

1 code implementation25 Sep 2017 Di Zhuang, J. Morris Chang, Mingchen Li

Community detection is of great importance for online social network analysis.

Social and Information Networks Cryptography and Security

MedChatZH: a Better Medical Adviser Learns from Better Instructions

1 code implementation3 Sep 2023 Yang Tan, Mingchen Li, Zijie Huang, Huiqun Yu, Guisheng Fan

Generative large language models (LLMs) have shown great success in various applications, including question-answering (QA) and dialogue systems.

Question Answering

AutoBalance: Optimized Loss Functions for Imbalanced Data

1 code implementation NeurIPS 2021 Mingchen Li, Xuechen Zhang, Christos Thrampoulidis, Jiasi Chen, Samet Oymak

Our experimental findings are complemented with theoretical insights on loss function design and the benefits of train-validation split.

Data Augmentation Fairness

Locally Differentially Private Distributed Deep Learning via Knowledge Distillation

1 code implementation7 Feb 2022 Di Zhuang, Mingchen Li, J. Morris Chang

As such, it motivates the researchers to conduct distributed deep learning, where the data user would like to build DL models using the data segregated across multiple different data owners.

Knowledge Distillation Privacy Preserving

Provable and Efficient Continual Representation Learning

1 code implementation3 Mar 2022 Yingcong Li, Mingchen Li, M. Salman Asif, Samet Oymak

In continual learning (CL), the goal is to design models that can learn a sequence of tasks without catastrophic forgetting.

Continual Learning Representation Learning

A Hierarchical N-Gram Framework for Zero-Shot Link Prediction

1 code implementation16 Apr 2022 Mingchen Li, Junfan Chen, Samuel Mensah, Nikolaos Aletras, Xiulong Yang, Yang Ye

Thus, in this paper, we propose a Hierarchical N-Gram framework for Zero-Shot Link Prediction (HNZSLP), which considers the dependencies among character n-grams of the relation surface name for ZSLP.

Knowledge Graphs Link Prediction +1

Federated Multi-Sequence Stochastic Approximation with Local Hypergradient Estimation

1 code implementation2 Jun 2023 Davoud Ataee Tarzanagh, Mingchen Li, Pranay Sharma, Samet Oymak

Stochastic approximation with multiple coupled sequences (MSA) has found broad applications in machine learning as it encompasses a rich class of problems including bilevel optimization (BLO), multi-level compositional optimization (MCO), and reinforcement learning (specifically, actor-critic methods).

Bilevel Optimization

Gradient Descent with Early Stopping is Provably Robust to Label Noise for Overparameterized Neural Networks

1 code implementation27 Mar 2019 Mingchen Li, Mahdi Soltanolkotabi, Samet Oymak

In particular, we prove that: (i) In the first few iterations where the updates are still in the vicinity of the initialization gradient descent only fits to the correct labels essentially ignoring the noisy labels.

Generalization Guarantees for Neural Networks via Harnessing the Low-rank Structure of the Jacobian

no code implementations12 Jun 2019 Samet Oymak, Zalan Fabian, Mingchen Li, Mahdi Soltanolkotabi

We show that over the information space learning is fast and one can quickly train a model with zero training loss that can also generalize well.

Multi-Fusion Chinese WordNet (MCW) : Compound of Machine Learning and Manual Correction

no code implementations5 Feb 2020 Mingchen Li, Zili Zhou, Yanna Wang

By using them, we found that these word networks have low accuracy and coverage, and cannot completely portray the semantic network of PWN.

BIG-bench Machine Learning Word Sense Disambiguation +1

Exploring Weight Importance and Hessian Bias in Model Pruning

no code implementations19 Jun 2020 Mingchen Li, Yahya Sattar, Christos Thrampoulidis, Samet Oymak

Model pruning is an essential procedure for building compact and computationally-efficient machine learning models.

On the Marginal Benefit of Active Learning: Does Self-Supervision Eat Its Cake?

no code implementations16 Nov 2020 Yao-Chun Chan, Mingchen Li, Samet Oymak

In parallel, recent developments in self-supervised and semi-supervised learning (S4L) provide powerful techniques, based on data-augmentation, contrastive learning, and self-training, that enable superior utilization of unlabeled data which led to a significant reduction in required labeling in the standard machine learning benchmarks.

Active Learning Contrastive Learning +1

Generalization Guarantees for Neural Architecture Search with Train-Validation Split

no code implementations29 Apr 2021 Samet Oymak, Mingchen Li, Mahdi Soltanolkotabi

In this approach, it is common to use bilevel optimization where one optimizes the model weights over the training data (inner problem) and various hyperparameters such as the configuration of the architecture over the validation data (outer problem).

Bilevel Optimization Generalization Bounds +2

GENERALIZATION GUARANTEES FOR NEURAL NETS VIA HARNESSING THE LOW-RANKNESS OF JACOBIAN

no code implementations25 Sep 2019 Samet Oymak, Zalan Fabian, Mingchen Li, Mahdi Soltanolkotabi

We show that over the information space learning is fast and one can quickly train a model with zero training loss that can also generalize well.

MC-GEN:Multi-level Clustering for Private Synthetic Data Generation

1 code implementation28 May 2022 Mingchen Li, Di Zhuang, J. Morris Chang

MC-GEN applies multi-level clustering and differential private generative model to improve the utility of synthetic data.

BIG-bench Machine Learning Clustering +2

SESNet: sequence-structure feature-integrated deep learning method for data-efficient protein engineering

no code implementations29 Dec 2022 Mingchen Li, Liqi Kang, Yi Xiong, Yu Guang Wang, Guisheng Fan, Pan Tan, Liang Hong

Here, we develop SESNet, a supervised deep-learning model to predict the fitness for protein mutants by leveraging both sequence and structure information, and exploiting attention mechanism.

Data Augmentation

TemPL: A Novel Deep Learning Model for Zero-Shot Prediction of Protein Stability and Activity Based on Temperature-Guided Language Modeling

no code implementations7 Apr 2023 Pan Tan, Mingchen Li, Liang Zhang, Zhiqiang Hu, Liang Hong

We introduce TemPL, a novel deep learning approach for zero-shot prediction of protein stability and activity, harnessing temperature-guided language modeling.

Language Modelling

Understand the Dynamic World: An End-to-End Knowledge Informed Framework for Open Domain Entity State Tracking

no code implementations26 Apr 2023 Mingchen Li, Lifu Huang

Open domain entity state tracking aims to predict reasonable state changes of entities (i. e., [attribute] of [entity] was [before_state] and [after_state] afterwards) given the action descriptions.

Attribute

How far is Language Model from 100% Few-shot Named Entity Recognition in Medical Domain

no code implementations1 Jul 2023 Mingchen Li, Rui Zhang

Recent advancements in language models (LMs) have led to the emergence of powerful models such as Small LMs (e. g., T5) and Large LMs (e. g., GPT-4).

few-shot-ner Few-shot NER +4

FedYolo: Augmenting Federated Learning with Pretrained Transformers

no code implementations10 Jul 2023 Xuechen Zhang, Mingchen Li, Xiangyu Chang, Jiasi Chen, Amit K. Roy-Chowdhury, Ananda Theertha Suresh, Samet Oymak

These insights on scale and modularity motivate a new federated learning approach we call "You Only Load Once" (FedYolo): The clients load a full PTF model once and all future updates are accomplished through communication-efficient modules with limited catastrophic-forgetting, where each task is assigned to its own module.

Federated Learning

Epi-Curriculum: Episodic Curriculum Learning for Low-Resource Domain Adaptation in Neural Machine Translation

no code implementations6 Sep 2023 Keyu Chen, Di Zhuang, Mingchen Li, J. Morris Chang

Experiments on English-German and English-Romanian translation show that: (i) Epi-Curriculum improves both model's robustness and adaptability in seen and unseen domains; (ii) Our episodic training framework enhances the encoder and decoder's robustness to domain shift.

Domain Adaptation Machine Translation +2

Benchingmaking Large Langage Models in Biomedical Triple Extraction

no code implementations27 Oct 2023 Mingchen Li, Huixue Zhou, Rui Zhang

Biomedical triple extraction systems aim to automatically extract biomedical entities and relations between entities.

Language Modelling Large Language Model +3

A Review of Reinforcement Learning for Natural Language Processing, and Applications in Healthcare

no code implementations23 Oct 2023 Ying Liu, Haozhu Wang, Huixue Zhou, Mingchen Li, Yu Hou, Sicheng Zhou, Fang Wang, Rama Hoetzlein, Rui Zhang

It has gained significant attention in the field of Natural Language Processing (NLP) due to its ability to learn optimal strategies for tasks such as dialogue systems, machine translation, and question-answering.

Decision Making Machine Translation +5

Class-attribute Priors: Adapting Optimization to Heterogeneity and Fairness Objective

no code implementations25 Jan 2024 Xuechen Zhang, Mingchen Li, Jiasi Chen, Christos Thrampoulidis, Samet Oymak

Confirming this, under a gaussian mixture setting, we show that the optimal SVM classifier for balanced accuracy needs to be adaptive to the class attributes.

Attribute Fairness

A Condensed Transition Graph Framework for Zero-shot Link Prediction with Large Language Models

no code implementations16 Feb 2024 Mingchen Li, Chen Ling, Rui Zhang, Liang Zhao

To address this, in this work, we introduce a Condensed Transition Graph Framework for Zero-Shot Link Prediction (CTLP), which encodes all the paths' information in linear time complexity to predict unseen relations between entities, attaining both efficiency and information preservation.

Contrastive Learning Knowledge Graphs +1

Cannot find the paper you are looking for? You can Submit a new open access paper.