Text-based Person Retrieval
11 papers with code • 0 benchmarks • 0 datasets
Benchmarks
These leaderboards are used to track progress in Text-based Person Retrieval
Most implemented papers
PGDS: Pose-Guidance Deep Supervision for Mitigating Clothes-Changing in Person Re-Identification
Person Re-Identification (Re-ID) task seeks to enhance the tracking of multiple individuals by surveillance cameras.
DSSL: Deep Surroundings-person Separation Learning for Text-based Person Retrieval
Many previous methods on text-based person retrieval tasks are devoted to learning a latent common space mapping, with the purpose of extracting modality-invariant features from both visual and textual modality.
Text-based Person Search in Full Images via Semantic-Driven Proposal Generation
Finding target persons in full scene images with a query of text description has important practical applications in intelligent video surveillance. However, different from the real-world scenarios where the bounding boxes are not available, existing text-based person retrieval methods mainly focus on the cross modal matching between the query text descriptions and the gallery of cropped pedestrian images.
See Finer, See More: Implicit Modality Alignment for Text-based Person Retrieval
To explore the fine-grained alignment, we further propose two implicit semantic alignment paradigms: multi-level alignment (MLA) and bidirectional mask modeling (BMM).
A Simple and Robust Correlation Filtering Method for Text-based Person Search
Text-based person search aims to associate pedestrian images with natural language descriptions.
Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person Retrieval
To alleviate these issues, we present IRRA: a cross-modal Implicit Relation Reasoning and Aligning framework that learns relations between local visual-textual tokens and enhances global image-text matching without requiring additional prior supervision.
PLIP: Language-Image Pre-training for Person Representation Learning
To address this issue, we propose a novel language-image pre-training framework for person representation learning, termed PLIP.
Towards Unified Text-based Person Retrieval: A Large-scale Multi-Attribute and Language Search Benchmark
To verify the feasibility of learning from the generated data, we develop a new joint Attribute Prompt Learning and Text Matching Learning (APTM) framework, considering the shared knowledge between attribute and text.
CLIP-based Synergistic Knowledge Transfer for Text-based Person Retrieval
Text-based Person Retrieval (TPR) aims to retrieve the target person images given a textual query.
Word4Per: Zero-shot Composed Person Retrieval
Searching for specific person has great social benefits and security value, and it often involves a combination of visual and textual information.