Search Results for author: Yong Luo

Found 69 papers, 35 papers with code

Unsupervised Domain Adaptation Method with Semantic-Structural Alignment for Dependency Parsing

no code implementations Findings (EMNLP) 2021 Boda Lin, Mingzheng Li, Si Li, Yong Luo

Unsupervised cross-domain dependency parsing is to accomplish domain adaptation for dependency parsing without using labeled data in target domain.

Dependency Parsing Unsupervised Domain Adaptation

Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging

1 code implementation16 Jan 2025 Anke Tang, Enneng Yang, Li Shen, Yong Luo, Han Hu, Bo Du, DaCheng Tao

In this study, we propose a training-free projection-based continual merging method that processes models sequentially through orthogonal projections of weight matrices and adaptive scaling mechanisms.

Leveraging Metamemory Mechanisms for Enhanced Data-Free Code Generation in LLMs

no code implementations14 Jan 2025 Shuai Wang, Liang Ding, Yibing Zhan, Yong Luo, Zheng He, Dapeng Tao

Automated code generation using large language models (LLMs) has gained attention due to its efficiency and adaptability.

Code Generation HumanEval

MambaHSI: Spatial-Spectral Mamba for Hyperspectral Image Classification

1 code implementation9 Jan 2025 Yapeng Li, Yong Luo, Lefei Zhang, Zengmao Wang, Bo Du

To remedy these drawbacks, we propose a novel HSI classification model based on a Mamba model, named MambaHSI, which can simultaneously model long-range interaction of the whole image and integrate spatial and spectral information in an adaptive manner.

Classification Hyperspectral Image Classification +1

Aligning Few-Step Diffusion Models with Dense Reward Difference Learning

1 code implementation18 Nov 2024 Ziyi Zhang, Li Shen, Sen Zhang, Deheng Ye, Yong Luo, Miaojing Shi, Bo Du, DaCheng Tao

Experimental results demonstrate that SDPO consistently outperforms prior methods in reward-based alignment across diverse step configurations, underscoring its robust step generalization capabilities.

Denoising

Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging

no code implementations29 Oct 2024 Li Shen, Anke Tang, Enneng Yang, Guibing Guo, Yong Luo, Lefei Zhang, Xiaochun Cao, Bo Du, DaCheng Tao

Building on WEMoE, we further introduce an efficient-and-effective WEMoE (E-WEMoE) method, whose core mechanism involves eliminating non-essential elements in the critical modules of WEMoE and implementing shared routing across multiple MoE modules, thereby significantly reducing both the trainable parameters, the overall parameter count, and computational overhead of the merged model by WEMoE.

Task Arithmetic

ADEM-VL: Adaptive and Embedded Fusion for Efficient Vision-Language Tuning

1 code implementation23 Oct 2024 Zhiwei Hao, Jianyuan Guo, Li Shen, Yong Luo, Han Hu, Yonggang Wen

To bridge this gap, we propose ADEM-VL, an efficient vision-language method that tunes VL models based on pretrained large language models (LLMs) by adopting a parameter-free cross-attention mechanism for similarity measurements in multimodal fusion.

Image Captioning Instruction Following +3

MG-Net: Learn to Customize QAOA with Circuit Depth Awareness

1 code implementation27 Sep 2024 Yang Qian, Xinbiao Wang, Yuxuan Du, Yong Luo, DaCheng Tao

To address this dilemma, here we first analyze the convergence behavior of QAOA, uncovering the origins of this dilemma and elucidating the intricate relationship between the employed mixer Hamiltonian, the specific problem at hand, and the permissible maximum circuit depth.

Combinatorial Optimization

$\mathbb{USCD}$: Improving Code Generation of LLMs by Uncertainty-Aware Selective Contrastive Decoding

no code implementations9 Sep 2024 Shuai Wang, Liang Ding, Li Shen, Yong Luo, Zheng He, Wei Yu, DaCheng Tao

Then, we selectively eliminate output noise induced by lame prompts based on the uncertainty of the prediction distribution from the standard prompt.

Code Generation HumanEval

Joint Input and Output Coordination for Class-Incremental Learning

no code implementations9 Sep 2024 Shuai Wang, Yibing Zhan, Yong Luo, Han Hu, Wei Yu, Yonggang Wen, DaCheng Tao

This mechanism assigns different weights to different categories of data according to the gradient of the output score, and uses knowledge distillation (KD) to reduce the mutual interference between the outputs of old and new tasks.

class-incremental learning Class Incremental Learning +2

Divide, Conquer and Combine: A Training-Free Framework for High-Resolution Image Perception in Multimodal Large Language Models

1 code implementation28 Aug 2024 Wenbin Wang, Liang Ding, Minyan Zeng, Xiabin Zhou, Li Shen, Yong Luo, DaCheng Tao

Building upon this insight, we propose Divide, Conquer and Combine (DC$^2$), a novel training-free framework for enhancing MLLM perception of HR images.

2k 4k +1

Sequential Federated Learning in Hierarchical Architecture on Non-IID Datasets

no code implementations19 Aug 2024 Xingrun Yan, Shiyuan Zuo, Rongfei Fan, Han Hu, Li Shen, Puning Zhao, Yong Luo

In a real federated learning (FL) system, communication overhead for passing model parameters between the clients and the parameter server (PS) is often a bottleneck.

Federated Learning

SMILE: Zero-Shot Sparse Mixture of Low-Rank Experts Construction From Pre-Trained Foundation Models

1 code implementation19 Aug 2024 Anke Tang, Li Shen, Yong Luo, Shuai Xie, Han Hu, Lefei Zhang, Bo Du, DaCheng Tao

Deep model training on extensive datasets is increasingly becoming cost-prohibitive, prompting the widespread adoption of deep model fusion techniques to leverage knowledge from pre-existing models.

Image Classification Text Generation

Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion

1 code implementation14 Jun 2024 Anke Tang, Li Shen, Yong Luo, Shiwei Liu, Han Hu, Bo Du

Once the routers are learned and a preference vector is set, the MoE module can be unloaded, thus no additional computational cost is introduced during inference.

Multi-Task Learning

FusionBench: A Comprehensive Benchmark of Deep Model Fusion

1 code implementation5 Jun 2024 Anke Tang, Li Shen, Yong Luo, Han Hu, Bo Du, DaCheng Tao

These techniques range from model ensemble methods, which combine the predictions to improve the overall performance, to model merging, which integrates different models into a single one, and model mixing methods, which upscale or recombine the components of the original models.

Image Classification model +4

Separable Power of Classical and Quantum Learning Protocols Through the Lens of No-Free-Lunch Theorem

no code implementations12 May 2024 Xinbiao Wang, Yuxuan Du, Kecheng Liu, Yong Luo, Bo Du, DaCheng Tao

The No-Free-Lunch (NFL) theorem, which quantifies problem- and data-independent generalization errors regardless of the optimization process, provides a foundational framework for comprehending diverse learning protocols' potential.

Attribute Quantum Machine Learning

Federated Learning with Only Positive Labels by Exploring Label Correlations

no code implementations24 Apr 2024 Xuming An, Dui Wang, Li Shen, Yong Luo, Han Hu, Bo Du, Yonggang Wen, DaCheng Tao

Specifically, FedALC estimates the label correlations in the class embedding learning for different label pairs and utilizes it to improve the model training.

Federated Learning Multi-Label Classification +1

MTGA: Multi-View Temporal Granularity Aligned Aggregation for Event-Based Lip-Reading

1 code implementation18 Apr 2024 WenHao Zhang, Jun Wang, Yong Luo, Lei Yu, Wei Yu, Zheng He, Jialie Shen

Then we design a spatio-temporal fusion module based on temporal granularity alignment, where the global spatial features extracted from event frames, together with the local relative spatial and temporal features contained in voxel graph list are effectively aligned and integrated.

Lip Reading

Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases

1 code implementation13 Feb 2024 Ziyi Zhang, Sen Zhang, Yibing Zhan, Yong Luo, Yonggang Wen, DaCheng Tao

Then, we surprisingly discover that dormant neurons in our critic model act as a regularization against reward overoptimization while active neurons reflect primacy bias.

Denoising Inductive Bias

Merging Multi-Task Models via Weight-Ensembling Mixture of Experts

1 code implementation1 Feb 2024 Anke Tang, Li Shen, Yong Luo, Nan Yin, Lefei Zhang, DaCheng Tao

A notable challenge is mitigating the interference between parameters of different models, which can substantially deteriorate performance.

Task Arithmetic

OOP: Object-Oriented Programming Evaluation Benchmark for Large Language Models

1 code implementation12 Jan 2024 Shuai Wang, Liang Ding, Li Shen, Yong Luo, Bo Du, DaCheng Tao

Advancing automated programming necessitates robust and comprehensive code generation benchmarks, yet current evaluation frameworks largely neglect object-oriented programming (OOP) in favor of functional programming (FP), e. g., HumanEval and MBPP.

Code Generation HumanEval

Improving Generalized Zero-Shot Learning by Exploring the Diverse Semantics from External Class Names

1 code implementation CVPR 2024 Yapeng Li, Yong Luo, Zengmao Wang, Bo Du

This motivates us to study GZSL in the more practical setting where unseen classes can be either similar or dissimilar to seen classes.

Generalized Zero-Shot Learning Test unseen

Concrete Subspace Learning based Interference Elimination for Multi-task Model Fusion

1 code implementation11 Dec 2023 Anke Tang, Li Shen, Yong Luo, Liang Ding, Han Hu, Bo Du, DaCheng Tao

At the upper level, we focus on learning a shared Concrete mask to identify the subspace, while at the inner level, model merging is performed to maximize the performance of the merged model.

Meta-Learning Task Arithmetic

Learn From Model Beyond Fine-Tuning: A Survey

1 code implementation12 Oct 2023 Hongling Zheng, Li Shen, Anke Tang, Yong Luo, Han Hu, Bo Du, DaCheng Tao

LFM focuses on the research, modification, and design of FM based on the model interface, so as to better understand the model structure and weights (in a black box environment), and to generalize the model to downstream tasks.

Meta-Learning model +2

Parameter Efficient Multi-task Model Fusion with Partial Linearization

1 code implementation7 Oct 2023 Anke Tang, Li Shen, Yong Luo, Yibing Zhan, Han Hu, Bo Du, Yixin Chen, DaCheng Tao

We demonstrate that our partial linearization technique enables a more effective fusion of multiple tasks into a single model, outperforming standard adapter tuning and task arithmetic alone.

parameter-efficient fine-tuning Task Arithmetic

Decompose Semantic Shifts for Composed Image Retrieval

no code implementations18 Sep 2023 Xingyu Yang, Daqing Liu, Heng Zhang, Yong Luo, Chaoyue Wang, Jing Zhang

Composed image retrieval is a type of image retrieval task where the user provides a reference image as a starting point and specifies a text on how to shift from the starting point to the desired target image.

Image Retrieval Retrieval

DeViT: Decomposing Vision Transformers for Collaborative Inference in Edge Devices

no code implementations10 Sep 2023 Guanyu Xu, Zhiwei Hao, Yong Luo, Han Hu, Jianping An, Shiwen Mao

Our objective is to achieve fast and energy-efficient collaborative inference while maintaining comparable accuracy compared with large ViTs.

Collaborative Inference Knowledge Distillation

PartSeg: Few-shot Part Segmentation via Part-aware Prompt Learning

no code implementations24 Aug 2023 Mengya Han, Heliang Zheng, Chaoyue Wang, Yong Luo, Han Hu, Jing Zhang, Yonggang Wen

In this work, we address the task of few-shot part segmentation, which aims to segment the different parts of an unseen object using very few labeled examples.

Language Modeling Language Modelling +1

Rethinking the Localization in Weakly Supervised Object Localization

no code implementations11 Aug 2023 Rui Xu, Yong Luo, Han Hu, Bo Du, Jialie Shen, Yonggang Wen

Weakly supervised object localization (WSOL) is one of the most popular and challenging tasks in computer vision.

Object Weakly-Supervised Object Localization

LGViT: Dynamic Early Exiting for Accelerating Vision Transformer

1 code implementation1 Aug 2023 Guanyu Xu, Jiawei Hao, Li Shen, Han Hu, Yong Luo, Hui Lin, Jialie Shen

Recently, the efficient deployment and acceleration of powerful vision transformers (ViTs) on resource-limited edge devices for providing multimedia services have become attractive tasks.

Multi-Granularity Hand Action Detection

2 code implementations19 Jun 2023 Ting Zhe, Jing Zhang, YongQian Li, Yong Luo, Han Hu, DaCheng Tao

To fill this gap, we introduce the FHA-Kitchens (Fine-Grained Hand Actions in Kitchen Scenes) dataset, providing both coarse- and fine-grained hand action categories along with localization annotations.

Action Detection Action Localization +6

Transition Role of Entangled Data in Quantum Machine Learning

1 code implementation6 Jun 2023 Xinbiao Wang, Yuxuan Du, Zhuozhuo Tu, Yong Luo, Xiao Yuan, DaCheng Tao

Recent progress has highlighted its positive impact on learning quantum dynamics, wherein the integration of entanglement into quantum operations or measurements of quantum machine learning (QML) models leads to substantial reductions in training data size, surpassing a specified prediction error threshold.

Quantum Machine Learning

Improving Heterogeneous Model Reuse by Density Estimation

1 code implementation23 May 2023 Anke Tang, Yong Luo, Han Hu, Fengxiang He, Kehua Su, Bo Du, Yixin Chen, DaCheng Tao

This paper studies multiparty learning, aiming to learn a model using the private data of different participants.

Density Estimation model +1

Unsupervised Cross-domain Pulmonary Nodule Detection without Source Data

1 code implementation3 Apr 2023 Rui Xu, Yong Luo, Bo Du

Cross-domain pulmonary nodule detection suffers from performance degradation due to a large shift of data distributions between the source and target domain.

Contrastive Learning object-detection +2

SGDA: Towards 3D Universal Pulmonary Nodule Detection via Slice Grouped Domain Attention

1 code implementation7 Mar 2023 Rui Xu, Zhi Liu, Yong Luo, Han Hu, Li Shen, Bo Du, Kaiming Kuang, Jiancheng Yang

To address this issue, we propose a slice grouped domain attention (SGDA) module to enhance the generalization capability of the pulmonary nodule detection networks.

Computed Tomography (CT)

FedABC: Targeting Fair Competition in Personalized Federated Learning

no code implementations15 Feb 2023 Dui Wang, Li Shen, Yong Luo, Han Hu, Kehua Su, Yonggang Wen, DaCheng Tao

In particular, we adopt the ``one-vs-all'' training strategy in each client to alleviate the unfair competition between classes by constructing a personalized binary classification problem for each class.

Binary Classification Personalized Federated Learning

MIGPerf: A Comprehensive Benchmark for Deep Learning Training and Inference Workloads on Multi-Instance GPUs

1 code implementation1 Jan 2023 Huaizheng Zhang, Yuanming Li, Wencong Xiao, Yizheng Huang, Xing Di, Jianxiong Yin, Simon See, Yong Luo, Chiew Tong Lau, Yang You

The vision of this paper is to provide a more comprehensive and practical benchmark study for MIG in order to eliminate the need for tedious manual benchmarking and tuning efforts.

Benchmarking

Depression Diagnosis and Analysis via Multimodal Multi-order Factor Fusion

no code implementations31 Dec 2022 Chengbo Yuan, Qianhui Xu, Yong Luo

Multimodal learning is a popular solution for automatic diagnosis of depression, and the existing works suffer two main drawbacks: 1) the high-order interactions between different modalities can not be well exploited; and 2) interpretability of the models are weak.

Not All Instances Contribute Equally: Instance-adaptive Class Representation Learning for Few-Shot Visual Recognition

no code implementations7 Sep 2022 Mengya Han, Yibing Zhan, Yong Luo, Bo Du, Han Hu, Yonggang Wen, DaCheng Tao

To address the above issues, we propose a novel metric-based meta-learning framework termed instance-adaptive class representation learning network (ICRL-Net) for few-shot visual recognition.

Meta-Learning Representation Learning

Symmetric Pruning in Quantum Neural Networks

no code implementations30 Aug 2022 Xinbiao Wang, Junyu Liu, Tongliang Liu, Yong Luo, Yuxuan Du, DaCheng Tao

To fill this knowledge gap, here we propose the effective quantum neural tangent kernel (EQNTK) and connect this concept with over-parameterization theory to quantify the convergence of QNNs towards the global optima.

LSSANet: A Long Short Slice-Aware Network for Pulmonary Nodule Detection

1 code implementation3 Aug 2022 Rui Xu, Yong Luo, Bo Du, Kaiming Kuang, Jiancheng Yang

Convolutional neural networks (CNNs) have been demonstrated to be highly effective in the field of pulmonary nodule detection.

Computed Tomography (CT)

Leveraging GAN Priors for Few-Shot Part Segmentation

1 code implementation27 Jul 2022 Mengya Han, Heliang Zheng, Chaoyue Wang, Yong Luo, Han Hu, Bo Du

Overall, this work is an attempt to explore the internal relevance between generation tasks and perception tasks by prompt designing.

Image Generation Segmentation

CLNode: Curriculum Learning for Node Classification

1 code implementation15 Jun 2022 Xiaowen Wei, Xiuwen Gong, Yibing Zhan, Bo Du, Yong Luo, Wenbin Hu

Experimental results on real-world networks demonstrate that CLNode is a general framework that can be combined with various GNNs to improve their accuracy and robustness.

Classification Node Classification

CDFKD-MFS: Collaborative Data-free Knowledge Distillation via Multi-level Feature Sharing

1 code implementation24 May 2022 Zhiwei Hao, Yong Luo, Zhi Wang, Han Hu, Jianping An

To tackle this challenge, we propose a framework termed collaborative data-free knowledge distillation via multi-level feature sharing (CDFKD-MFS), which consists of a multi-header student module, an asymmetric adversarial data-free KD module, and an attention-based aggregation module.

Data-free Knowledge Distillation

Multi-Agent Collaborative Inference via DNN Decoupling: Intermediate Feature Compression and Edge Learning

1 code implementation24 May 2022 Zhiwei Hao, Guanyu Xu, Yong Luo, Han Hu, Jianping An, Shiwen Mao

In this paper, we study the multi-agent collaborative inference scenario, where a single edge server coordinates the inference of multiple UEs.

Collaborative Inference Feature Compression

Unpaired Image Captioning by Image-level Weakly-Supervised Visual Concept Recognition

no code implementations7 Mar 2022 Peipei Zhu, Xiao Wang, Yong Luo, Zhenglong Sun, Wei-Shi Zheng, YaoWei Wang, Changwen Chen

The image-level labels are utilized to train a weakly-supervised object recognition model to extract object information (e. g., instance) in an image, and the extracted instances are adopted to infer the relationships among different objects based on an enhanced graph neural network (GNN).

Graph Neural Network Image Captioning +3

Hyper-relationship Learning Network for Scene Graph Generation

no code implementations15 Feb 2022 Yibing Zhan, Zhi Chen, Jun Yu, Baosheng Yu, DaCheng Tao, Yong Luo

As a result, HLN significantly improves the performance of scene graph generation by integrating and reasoning from object interactions, relationship interactions, and transitive inference of hyper-relationships.

Graph Attention Graph Generation +1

Schema-Free Dependency Parsing via Sequence Generation

no code implementations28 Jan 2022 Boda Lin, Zijun Yao, Jiaxin Shi, Shulin Cao, Binghao Tang, Si Li, Yong Luo, Juanzi Li, Lei Hou

To remedy these drawbacks, we propose to achieve universal and schema-free Dependency Parsing (DP) via Sequence Generation (SG) DPSG by utilizing only the pre-trained language model (PLM) without any auxiliary structures or parsing algorithms.

Decoder Dependency Parsing +1

Resistance Training using Prior Bias: toward Unbiased Scene Graph Generation

1 code implementation18 Jan 2022 Chao Chen, Yibing Zhan, Baosheng Yu, Liu Liu, Yong Luo, Bo Du

To address this problem, we propose Resistance Training using Prior Bias (RTPB) for the scene graph generation.

Graph Generation Unbiased Scene Graph Generation

ViF-SD2E: A Robust Weakly-Supervised Method for Neural Decoding

no code implementations2 Dec 2021 Jingyi Feng, Yong Luo, Shuang Song

Neural decoding plays a vital role in the interaction between the brain and the outside world.

ModelPS: An Interactive and Collaborative Platform for Editing Pre-trained Models at Scale

1 code implementation18 May 2021 Yuanming Li, Huaizheng Zhang, Shanshan Jiang, Fan Yang, Yonggang Wen, Yong Luo

AI engineering has emerged as a crucial discipline to democratize deep neural network (DNN) models among software developers with a diverse background.

Model Editing

Towards understanding the power of quantum kernels in the NISQ era

no code implementations31 Mar 2021 Xinbiao Wang, Yuxuan Du, Yong Luo, DaCheng Tao

In this study, we fill this knowledge gap by exploiting the power of quantum kernels when the quantum system noise and sample error are considered.

Open-Ended Question Answering Quantum Machine Learning

A Serverless Cloud-Fog Platform for DNN-Based Video Analytics with Incremental Learning

no code implementations5 Feb 2021 Huaizheng Zhang, Meng Shen, Yizheng Huang, Yonggang Wen, Yong Luo, Guanyu Gao, Kyle Guan

To save bandwidth and reduce RTT, VPaaS provides a new video streaming protocol that only sends low-quality video to the cloud.

Incremental Learning Management

Hysia: Serving DNN-Based Video-to-Retail Applications in Cloud

2 code implementations9 Jun 2020 Huaizheng Zhang, Yuanming Li, Qiming Ai, Yong Luo, Yonggang Wen, Yichao Jin, Nguyen Binh Duong Ta

Combining \underline{v}ideo streaming and online \underline{r}etailing (V2R) has been a growing trend recently.

Video Retrieval Video-to-Shop

Look, Read and Feel: Benchmarking Ads Understanding with Multimodal Multitask Learning

no code implementations21 Dec 2019 Huaizheng Zhang, Yong Luo, Qiming Ai, Yonggang Wen

A multitask loss function is also designed to train both the topic and sentiment prediction models jointly in an end-to-end manner.

Benchmarking

Towards Digital Retina in Smart Cities: A Model Generation, Utilization and Communication Paradigm

1 code implementation31 Jul 2019 Yihang Lou, Ling-Yu Duan, Yong Luo, Ziqian Chen, Tongliang Liu, Shiqi Wang, Wen Gao

The digital retina in smart cities is to select what the City Eye tells the City Brain, and convert the acquired visual data from front-end visual sensors to features in an intelligent sensing manner.

Deep Learning

Decomposition-Based Transfer Distance Metric Learning for Image Classification

no code implementations8 Apr 2019 Yong Luo, Tongliang Liu, DaCheng Tao, Chao Xu

In particular, DTDML learns a sparse combination of the base metrics to construct the target metric by forcing the target metric to be close to an integration of the source metrics.

Classification General Classification +3

Multi-view Vector-valued Manifold Regularization for Multi-label Image Classification

no code implementations8 Apr 2019 Yong Luo, DaCheng Tao, Chang Xu, Chao Xu, Hong Liu, Yonggang Wen

In computer vision, image datasets used for classification are naturally associated with multiple labels and comprised of multiple views, because each image may contain several objects (e. g. pedestrian, bicycle and tree) and is properly characterized by multiple visual features (e. g. color, texture and shape).

General Classification Multi-Label Image Classification

Transferring Knowledge Fragments for Learning Distance Metric from A Heterogeneous Domain

no code implementations8 Apr 2019 Yong Luo, Yonggang Wen, Tongliang Liu, DaCheng Tao

Some existing heterogeneous transfer learning (HTL) approaches can learn target distance metric by usually transforming the samples of source and target domain into a common subspace.

Metric Learning Transfer Learning

Multi-View Matrix Completion for Multi-Label Image Classification

no code implementations8 Apr 2019 Yong Luo, Tongliang Liu, DaCheng Tao, Chao Xu

Therefore, we propose to weightedly combine the MC outputs of different views, and present the multi-view matrix completion (MVMC) framework for transductive multi-label image classification.

Classification General Classification +6

Heterogeneous Multi-task Metric Learning across Multiple Domains

no code implementations8 Apr 2019 Yong Luo, Yonggang Wen, DaCheng Tao

Heterogeneous transfer learning approaches can be adopted to remedy this drawback by deriving a metric from the learned transformation across different domains.

Metric Learning Scene Classification +2

Cost-Sensitive Feature Selection by Optimizing F-Measures

no code implementations4 Apr 2019 Meng Liu, Chang Xu, Yong Luo, Chao Xu, Yonggang Wen, DaCheng Tao

Feature selection is beneficial for improving the performance of general machine learning tasks by extracting an informative subset from the high-dimensional features.

feature selection

Transfer Metric Learning: Algorithms, Applications and Outlooks

no code implementations9 Oct 2018 Yong Luo, Yonggang Wen, Ling-Yu Duan, DaCheng Tao

Distance metric learning (DML) aims to find an appropriate way to reveal the underlying data relationship.

Metric Learning Triplet

Tensor Canonical Correlation Analysis for Multi-view Dimension Reduction

3 code implementations9 Feb 2015 Yong Luo, DaCheng Tao, Yonggang Wen, Kotagiri Ramamohanarao, Chao Xu

As a consequence, the high order correlation information contained in the different views is explored and thus a more reliable common subspace shared by all features can be obtained.

Dimensionality Reduction MULTI-VIEW LEARNING

Cannot find the paper you are looking for? You can Submit a new open access paper.