Search Results for author: Shi Yu

Found 19 papers, 11 papers with code

Say More with Less: Understanding Prompt Learning Behaviors through Gist Compression

1 code implementation25 Feb 2024 Xinze Li, Zhenghao Liu, Chenyan Xiong, Shi Yu, Yukun Yan, Shuo Wang, Ge Yu

It finetunes the compression plugin module and uses the representations of gist tokens to emulate the raw prompts in the vanilla language model.

Language Modelling

ActiveRAG: Revealing the Treasures of Knowledge via Active Learning

1 code implementation21 Feb 2024 Zhipeng Xu, Zhenghao Liu, Yibin Liu, Chenyan Xiong, Yukun Yan, Shuo Wang, Shi Yu, Zhiyuan Liu, Ge Yu

Retrieval Augmented Generation (RAG) has introduced a new paradigm for Large Language Models (LLMs), aiding in the resolution of knowledge-intensive tasks.

Active Learning Position +2

Text Matching Improves Sequential Recommendation by Reducing Popularity Biases

1 code implementation27 Aug 2023 Zhenghao Liu, Sen Mei, Chenyan Xiong, Xiaohua LI, Shi Yu, Zhiyuan Liu, Yu Gu, Ge Yu

TASTE alleviates the cold start problem by representing long-tail items using full-text modeling and bringing the benefits of pretrained language models to recommendation systems.

Sequential Recommendation Text Matching

Structure-Aware Language Model Pretraining Improves Dense Retrieval on Structured Data

1 code implementation31 May 2023 Xinze Li, Zhenghao Liu, Chenyan Xiong, Shi Yu, Yu Gu, Zhiyuan Liu, Ge Yu

SANTA proposes two pretraining methods to make language models structure-aware and learn effective representations for structured data: 1) Structured Data Alignment, which utilizes the natural alignment relations between structured data and unstructured data for structure-aware pretraining.

Code Search Language Modelling +1

Augmentation-Adapted Retriever Improves Generalization of Language Models as Generic Plug-In

1 code implementation27 May 2023 Zichun Yu, Chenyan Xiong, Shi Yu, Zhiyuan Liu

Retrieval augmentation can aid language models (LMs) in knowledge-intensive tasks by supplying them with external information.

Retrieval Zero-shot Generalization

Fusion-in-T5: Unifying Document Ranking Signals for Improved Information Retrieval

no code implementations24 May 2023 Shi Yu, Chenghao Fan, Chenyan Xiong, David Jin, Zhiyuan Liu, Zhenghao Liu

Common IR pipelines are typically cascade systems that may involve multiple rankers and/or fusion models to integrate different information step-by-step.

Document Ranking Information Retrieval +2

Rethinking Dense Retrieval's Few-Shot Ability

1 code implementation12 Apr 2023 Si Sun, Yida Lu, Shi Yu, Xiangyang Li, Zhonghua Li, Zhao Cao, Zhiyuan Liu, Deiming Ye, Jie Bao

Moreover, the dataset is disjointed into base and novel classes, allowing DR models to be continuously trained on ample data from base classes and a few samples in novel classes.

Retrieval

P^3 Ranker: Mitigating the Gaps between Pre-training and Ranking Fine-tuning with Prompt-based Learning and Pre-finetuning

1 code implementation4 May 2022 Xiaomeng Hu, Shi Yu, Chenyan Xiong, Zhenghao Liu, Zhiyuan Liu, Ge Yu

In this paper, we identify and study the two mismatches between pre-training and ranking fine-tuning: the training schema gap regarding the differences in training objectives and model architectures, and the task knowledge gap considering the discrepancy between the knowledge needed in ranking and that learned during pre-training.

Robo-Advising: Enhancing Investment with Inverse Optimization and Deep Reinforcement Learning

no code implementations19 May 2021 Haoran Wang, Shi Yu

Machine Learning (ML) has been embraced as a powerful tool by the financial industry, with notable applications spreading in various domains including investment management.

Management Portfolio Optimization +2

Few-Shot Conversational Dense Retrieval

1 code implementation10 May 2021 Shi Yu, Zhenghao Liu, Chenyan Xiong, Tao Feng, Zhiyuan Liu

In this paper, we present a Conversational Dense Retrieval system, ConvDR, that learns contextualized embeddings for multi-turn conversational queries and retrieves documents solely using embedding dot products.

Conversational Search Retrieval

Exploring Fluent Query Reformulations with Text-to-Text Transformers and Reinforcement Learning

no code implementations18 Dec 2020 Jerry Zikun Chen, Shi Yu, Haoran Wang

Query reformulation aims to alter noisy or ambiguous text sequences into coherent ones closer to natural language questions.

intent-classification Intent Classification +4

Learning Risk Preferences from Investment Portfolios Using Inverse Optimization

no code implementations4 Oct 2020 Shi Yu, Haoran Wang, Chaosheng Dong

Our approach allows the learner to continuously estimate real-time risk preferences using concurrent observed portfolios and market price data.

Decision Making Management

Few-Shot Generative Conversational Query Rewriting

1 code implementation9 Jun 2020 Shi Yu, Jiahua Liu, Jingqin Yang, Chenyan Xiong, Paul Bennett, Jianfeng Gao, Zhiyuan Liu

Conversational query rewriting aims to reformulate a concise conversational query to a fully specified, context-independent query that can be effectively handled by existing information retrieval systems.

Information Retrieval Retrieval +2

Eigendecomposition of Q in Equally Constrained Quadratic Programming

1 code implementation22 Apr 2020 Shi Yu

When applying eigenvalue decomposition on the quadratic term matrix in a type of linear equally constrained quadratic programming (EQP), there exists a linear mapping to project optimal solutions between the new EQP formulation where $Q$ is diagonalized and the original formulation.

A Financial Service Chatbot based on Deep Bidirectional Transformers

no code implementations17 Feb 2020 Shi Yu, Yuxin Chen, Hussain Zaidi

Our main novel contribution is the discussion about uncertainty measure for BERT, where three different approaches are systematically compared on real problems.

Chatbot intent-classification +3

Cannot find the paper you are looking for? You can Submit a new open access paper.