Search Results for author: Jingtao Zhan

Found 12 papers, 10 papers with code

Capability-aware Prompt Reformulation Learning for Text-to-Image Generation

1 code implementation27 Mar 2024 Jingtao Zhan, Qingyao Ai, Yiqun Liu, Jia Chen, Shaoping Ma

Our in-depth analysis of these logs reveals that user prompt reformulation is heavily dependent on the individual user's capability, resulting in significant variance in the quality of reformulation pairs.

Text-to-Image Generation

Scaling Laws For Dense Retrieval

no code implementations27 Mar 2024 Yan Fang, Jingtao Zhan, Qingyao Ai, Jiaxin Mao, Weihang Su, Jia Chen, Yiqun Liu

In this study, we investigate whether the performance of dense retrieval models follows the scaling law as other neural models.

Data Augmentation Retrieval +1

Query Augmentation by Decoding Semantics from Brain Signals

1 code implementation24 Feb 2024 Ziyi Ye, Jingtao Zhan, Qingyao Ai, Yiqun Liu, Maarten de Rijke, Christina Lioma, Tuukka Ruotsalo

If the quality of the initially retrieved documents is low, then the effectiveness of query augmentation would be limited as well.

Document Ranking

Constructing Tree-based Index for Efficient and Effective Dense Retrieval

1 code implementation24 Apr 2023 Haitao Li, Qingyao Ai, Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Zheng Liu, Zhao Cao

Unfortunately, while ANN can improve the efficiency of DR models, it usually comes with a significant price on retrieval performance.

Contrastive Learning Retrieval

Disentangled Modeling of Domain and Relevance for Adaptable Dense Retrieval

1 code implementation11 Aug 2022 Jingtao Zhan, Qingyao Ai, Yiqun Liu, Jiaxin Mao, Xiaohui Xie, Min Zhang, Shaoping Ma

By making the REM and DAMs disentangled, DDR enables a flexible training paradigm in which REM is trained with supervision once and DAMs are trained with unsupervised data.

Ad-Hoc Information Retrieval Domain Adaptation +1

Evaluating Interpolation and Extrapolation Performance of Neural Retrieval Models

1 code implementation25 Apr 2022 Jingtao Zhan, Xiaohui Xie, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

For example, representation-based retrieval models perform almost as well as interaction-based retrieval models in terms of interpolation but not extrapolation.

Retrieval

Interpreting Dense Retrieval as Mixture of Topics

no code implementations27 Nov 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

Dense Retrieval (DR) reaches state-of-the-art results in first-stage retrieval, but little is known about the mechanisms that contribute to its success.

Retrieval

Learning Discrete Representations via Constrained Clustering for Effective and Efficient Dense Retrieval

4 code implementations12 Oct 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

However, the efficiency of most existing DR models is limited by the large memory cost of storing dense vectors and the time-consuming nearest neighbor search (NNS) in vector space.

Constrained Clustering Information Retrieval +3

Jointly Optimizing Query Encoder and Product Quantization to Improve Retrieval Performance

5 code implementations2 Aug 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

Compared with previous DR models that use brute-force search, JPQ almost matches the best retrieval performance with 30x compression on index size.

Information Retrieval Quantization +1

Optimizing Dense Retrieval Model Training with Hard Negatives

4 code implementations16 Apr 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

ADORE replaces the widely-adopted static hard negative sampling method with a dynamic one to directly optimize the ranking performance.

Information Retrieval Representation Learning +1

Learning To Retrieve: How to Train a Dense Retrieval Model Effectively and Efficiently

2 code implementations20 Oct 2020 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma

Through this process, it teaches the DR model how to retrieve relevant documents from the entire corpus instead of how to rerank a potentially biased sample of documents.

Passage Retrieval Retrieval

RepBERT: Contextualized Text Embeddings for First-Stage Retrieval

3 code implementations28 Jun 2020 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma

Although exact term match between queries and documents is the dominant method to perform first-stage retrieval, we propose a different approach, called RepBERT, to represent documents and queries with fixed-length contextualized embeddings.

Passage Ranking Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.