Search Results for author: Xiaopeng Li

Found 29 papers, 9 papers with code

DWE+: Dual-Way Matching Enhanced Framework for Multimodal Entity Linking

2 code implementations7 Apr 2024 Shezheng Song, Shasha Li, Shan Zhao, Xiaopeng Li, Chengyu Wang, Jie Yu, Jun Ma, Tianwei Yan, Bin Ji, Xiaoguang Mao

Multimodal entity linking (MEL) aims to utilize multimodal information (usually textual and visual information) to link ambiguous mentions to unambiguous entities in knowledge base.

Contrastive Learning Entity Linking

SWEA: Updating Factual Knowledge in Large Language Models via Subject Word Embedding Altering

1 code implementation31 Jan 2024 Xiaopeng Li, Shasha Li, Shezheng Song, Huijun Liu, Bin Ji, Xi Wang, Jun Ma, Jie Yu, Xiaodong Liu, Jing Wang, Weimin Zhang

In particular, local editing methods, which directly update model parameters, are more suitable for updating a small amount of knowledge.

Model Editing Word Embeddings

Agent4Ranking: Semantic Robust Ranking via Personalized Query Rewriting Using Multi-agent LLM

no code implementations24 Dec 2023 Xiaopeng Li, Lixin Su, Pengyue Jia, Xiangyu Zhao, Suqi Cheng, Junfeng Wang, Dawei Yin

To be specific, we use Chain of Thought (CoT) technology to utilize Large Language Models (LLMs) as agents to emulate various demographic profiles, then use them for efficient query rewriting, and we innovate a robust Multi-gate Mixture of Experts (MMoE) architecture coupled with a hybrid loss function, collectively strengthening the ranking models' robustness.

How to Bridge the Gap between Modalities: A Comprehensive Survey on Multimodal Large Language Model

no code implementations10 Nov 2023 Shezheng Song, Xiaopeng Li, Shasha Li, Shan Zhao, Jie Yu, Jun Ma, Xiaoguang Mao, Weimin Zhang

The study surveys existing modal alignment methods in MLLMs into four groups: (1) Multimodal Converters that change data into something LLMs can understand; (2) Multimodal Perceivers to improve how LLMs perceive different types of data; (3) Tools Assistance for changing data into one common format, usually text; and (4) Data-Driven methods that teach LLMs to understand specific types of data in a dataset.

Language Modelling Large Language Model

P-ROCKET: Pruning Random Convolution Kernels for Time Series Classification from a Feature Selection Perspective

1 code implementation15 Sep 2023 Shaowu Chen, Weize Sun, Lei Huang, Xiaopeng Li, Qingyuan Wang, Deepu John

In recent years, two competitive time series classification models, namely, ROCKET and MINIROCKET, have garnered considerable attention due to their low training cost and high accuracy.

Evolutionary Algorithms feature selection +2

HAMUR: Hyper Adapter for Multi-Domain Recommendation

1 code implementation12 Sep 2023 Xiaopeng Li, Fan Yan, Xiangyu Zhao, Yichao Wang, Bo Chen, Huifeng Guo, Ruiming Tang

Secondly, due to the distribution differences among domains, the utilization of static parameters in existing methods limits their flexibility to adapt to diverse domains.

Scenario-Aware Hierarchical Dynamic Network for Multi-Scenario Recommendation

no code implementations5 Sep 2023 Jingtong Gao, Bo Chen, Menghui Zhu, Xiangyu Zhao, Xiaopeng Li, Yuhao Wang, Yichao Wang, Huifeng Guo, Ruiming Tang

To address these limitations, we propose a Scenario-Aware Hierarchical Dynamic Network for Multi-Scenario Recommendations (HierRec), which perceives implicit patterns adaptively and conducts explicit and implicit scenario modeling jointly.

Click-Through Rate Prediction

Graph-Based Interaction-Aware Multimodal 2D Vehicle Trajectory Prediction using Diffusion Graph Convolutional Networks

no code implementations5 Sep 2023 Keshu Wu, Yang Zhou, Haotian Shi, Xiaopeng Li, Bin Ran

Within this framework, vehicles' motions are conceptualized as nodes in a time-varying graph, and the traffic interactions are represented by a dynamic adjacency matrix.

Graph Embedding Intent Detection +1

PMET: Precise Model Editing in a Transformer

1 code implementation17 Aug 2023 Xiaopeng Li, Shasha Li, Shezheng Song, Jing Yang, Jun Ma, Jie Yu

To achieve more precise model editing, we analyze hidden states of MHSA and FFN, finding that MHSA encodes certain general knowledge extraction patterns.

General Knowledge Model Editing

Multi-lingual Evaluation of Code Generation Models

2 code implementations26 Oct 2022 Ben Athiwaratkun, Sanjay Krishna Gouda, Zijian Wang, Xiaopeng Li, Yuchen Tian, Ming Tan, Wasi Uddin Ahmad, Shiqi Wang, Qing Sun, Mingyue Shang, Sujan Kumar Gonugondla, Hantian Ding, Varun Kumar, Nathan Fulton, Arash Farahani, Siddhartha Jain, Robert Giaquinto, Haifeng Qian, Murali Krishna Ramanathan, Ramesh Nallapati, Baishakhi Ray, Parminder Bhatia, Sudipta Sengupta, Dan Roth, Bing Xiang

Using these benchmarks, we are able to assess the performance of code generation models in a multi-lingual fashion, and discovered generalization ability of language models on out-of-domain languages, advantages of multi-lingual models over mono-lingual, the ability of few-shot prompting to teach the model new languages, and zero-shot translation abilities even on mono-lingual settings.

Code Completion Code Translation +1

Hard instance learning for quantum adiabatic prime factorization

no code implementations10 Oct 2021 Jian Lin, Zhengfeng Zhang, Junping Zhang, Xiaopeng Li

Prime factorization is a difficult problem with classical computing, whose exponential hardness is the foundation of Rivest-Shamir-Adleman (RSA) cryptography.

Reinforcement Learning (RL) Transfer Learning

Uncertainty Set Prediction of Aggregated Wind Power Generation based on Bayesian LSTM and Spatio-Temporal Analysis

no code implementations7 Oct 2021 Xiaopeng Li, Jiang Wu, Zhanbo Xu, Kun Liu, Jun Yu, Xiaohong Guan

This paper focuses on the uncertainty set prediction of the aggregated generation of geographically distributed wind farms.

Quantum Adiabatic Doping for Atomic Fermi-Hubbard Quantum Simulations

no code implementations5 Jan 2021 Jue Nan, Jian Lin, Yuchen Luo, Bo Zhao, Xiaopeng Li

Its feasibility has been demonstrated with numerical simulations of the adiabatic preparation for certain incommensurate particle-doping fractions, where the major problem to circumvent is the atomic localization in the incommensurate lattice.

Quantum Gases Strongly Correlated Electrons Quantum Physics

Not All Attention Is Needed: Gated Attention Network for Sequence Data

1 code implementation1 Dec 2019 Lanqing Xue, Xiaopeng Li, Nevin L. Zhang

Attention mechanisms compute input-dependent dynamic attention weights for aggregating a sequence of hidden states.

Sentence text-classification +1

Learning to Abstract for Memory-augmented Conversational Response Generation

1 code implementation ACL 2019 Zhiliang Tian, Wei Bi, Xiaopeng Li, Nevin L. Zhang

In this work, we propose a memory-augmented generative model, which learns to abstract from the training corpus and saves the useful information to the memory to assist the response generation.

Conversational Response Generation Informativeness +2

Quantum Adiabatic Algorithm Design using Reinforcement Learning

no code implementations27 Dec 2018 Jian Lin, Zhong Yuan Lai, Xiaopeng Li

We benchmark this approach in Grover-search and 3-SAT problems, and find that the adiabatic-algorithm obtained by our RL approach leads to significant improvement in the resultant success probability.

reinforcement-learning Reinforcement Learning (RL)

Review Helpfulness Assessment based on Convolutional Neural Network

no code implementations27 Aug 2018 Xianshan Qu, Xiaopeng Li, John R. Rose

In this paper we describe the implementation of a convolutional neural network (CNN) used to assess online review helpfulness.

Neural Machine Translation Inspired Binary Code Similarity Comparison beyond Function Pairs

no code implementations8 Aug 2018 Fei Zuo, Xiaopeng Li, Patrick Young, Lannan Luo, Qiang Zeng, Zhexin Zhang

The solutions to these two problems have many applications, such as cross-architecture vulnerability discovery and code plagiarism detection.

Machine Translation Translation

Learning Sparse Deep Feedforward Networks via Tree Skeleton Expansion

no code implementations16 Mar 2018 Zhourong Chen, Xiaopeng Li, Nevin L. Zhang

An important characteristic of FNN structures learned this way is that they are sparse.

Building Sparse Deep Feedforward Networks using Tree Receptive Fields

no code implementations14 Mar 2018 Xiaopeng Li, Zhourong Chen, Nevin L. Zhang

We use Chow-Liu's algorithm to learn a tree-structured probabilistic model for the units at the current level, use the tree to identify subsets of units that are strongly correlated, and introduce a new unit with receptive field over the subsets.

Learning Latent Superstructures in Variational Autoencoders for Deep Multidimensional Clustering

no code implementations ICLR 2019 Xiaopeng Li, Zhourong Chen, Leonard K. M. Poon, Nevin L. Zhang

We investigate a variant of variational autoencoders where there is a superstructure of discrete latent variables on top of the latent features.

Clustering

Learning Parsimonious Deep Feed-forward Networks

no code implementations ICLR 2018 Zhourong Chen, Xiaopeng Li, Nevin L. Zhang

Convolutional neural networks and recurrent neural networks are designed with network structures well suited to the nature of spacial and sequential data respectively.

Cannot find the paper you are looking for? You can Submit a new open access paper.