Search Results for author: Lan-Zhe Guo

Found 28 papers, 8 papers with code

Micro Text Classification Based on Balanced Positive-Unlabeled Learning

1 code implementation17 Mar 2025 Lin-Han Jia, Lan-Zhe Guo, Zhi Zhou, Si-Ye Han, Zi-Wen Li, Yu-Feng Li

In real-world text classification tasks, negative texts often contain a minimal proportion of negative content, which is especially problematic in areas like text quality control, legal risk screening, and sensitive information interception.

text-classification Text Classification

D3: Diversity, Difficulty, and Dependability-Aware Data Selection for Sample-Efficient LLM Instruction Tuning

no code implementations14 Mar 2025 Jia Zhang, Chen-Xi Zhang, Yao Liu, Yi-Xuan Jin, Xiao-Wen Yang, Bo Zheng, Yi Liu, Lan-Zhe Guo

In this paper, we first establish data selection criteria based on three distinct aspects of data value: diversity, difficulty, and dependability, and then propose the D3 method comprising two key steps of scoring and selection.

Diversity Instruction Following

LawGPT: Knowledge-Guided Data Generation and Its Application to Legal LLM

1 code implementation10 Feb 2025 Zhi Zhou, Kun-Yang Yu, Shi-Yu Tian, Jiang-Xin Shi, Xiao-Wen Yang, Pengxiao Song, Yi-Xuan Jin, Lan-Zhe Guo, Yu-Feng Li

To address these limitations, we study data generation for legal reasoning to improve the legal reasoning performance of open-source LLMs with the help of proprietary LLMs.

Legal Reasoning

Step Back to Leap Forward: Self-Backtracking for Boosting Reasoning of Language Models

1 code implementation6 Feb 2025 Xiao-Wen Yang, Xuan-Yi Zhu, Wen-Da Wei, Ding-Chu Zhang, Jie-Jing Shao, Zhi Zhou, Lan-Zhe Guo, Yu-Feng Li

The integration of slow-thinking mechanisms into large language models (LLMs) offers a promising way toward achieving Level 2 AGI Reasoners, as exemplified by systems like OpenAI's o1.

Bridging Internal Probability and Self-Consistency for Effective and Efficient LLM Reasoning

no code implementations1 Feb 2025 Zhi Zhou, Tan Yuhao, Zenan Li, Yuan YAO, Lan-Zhe Guo, Xiaoxing Ma, Yu-Feng Li

In this paper, we present the first theoretical error decomposition analysis of these techniques, breaking down their error into estimation error and model error.

TabFSBench: Tabular Benchmark for Feature Shifts in Open Environment

1 code implementation31 Jan 2025 Zi-Jian Cheng, Zi-Yi Jia, Zhi Zhou, Lan-Zhe Guo, Yu-Feng Li

TabFSBench evaluates impacts of four distinct feature-shift scenarios on four tabular model categories across various datasets and assesses the performance of large language models (LLMs) and tabular LLMs in the tabular benchmark for the first time.

Contrast-Aware Calibration for Fine-Tuned CLIP: Leveraging Image-Text Alignment

no code implementations31 Jan 2025 Song-Lin Lv, Yu-Yang Chen, Zhi Zhou, Yu-Feng Li, Lan-Zhe Guo

Vision-language models (VLMs), such as CLIP, have demonstrated exceptional generalization capabilities and can quickly adapt to downstream tasks through prompt fine-tuning.

Vision-Language Model Selection and Reuse for Downstream Adaptation

no code implementations30 Jan 2025 Hao-Zhe Tan, Zhi Zhou, Yu-Feng Li, Lan-Zhe Guo

The proposal is highly computationally efficient and growable since the model labeling process is completed target task independent and the ability could grow with the number of candidate VLMs.

Language Modeling Language Modelling +1

CGI: Identifying Conditional Generative Models with Example Images

no code implementations23 Jan 2025 Zhi Zhou, Hao-Zhe Tan, Peng-Xiao Song, Lan-Zhe Guo

In this paper, we propose Conditional Generative Model Identification (CGI), which aims to provide an effective way to identify the most suitable model using user-provided example images rather than requiring users to manually review a large number of models with example images.

Text Matching

BMIP: Bi-directional Modality Interaction Prompt Learning for VLM

no code implementations14 Jan 2025 Song-Lin Lv, Yu-Yang Chen, Zhi Zhou, Ming Yang, Lan-Zhe Guo

Vision-language models (VLMs) have exhibited remarkable generalization capabilities, and prompt learning for VLMs has attracted great attention for the ability to adapt pre-trained VLMs to specific downstream tasks.

Domain Generalization Prompt Learning

Robust Semi-Supervised Learning in Open Environments

no code implementations24 Dec 2024 Lan-Zhe Guo, Lin-Han Jia, Jie-Jing Shao, Yu-Feng Li

Conventional SSL studies typically assume close environments where important factors (e. g., label, feature, distribution) between labeled and unlabeled data are consistent.

ChinaTravel: A Real-World Benchmark for Language Agents in Chinese Travel Planning

no code implementations18 Dec 2024 Jie-Jing Shao, Xiao-Wen Yang, Bo-Wen Zhang, Baizhi Chen, Wen-Da Wei, Guohao Cai, Zhenhua Dong, Lan-Zhe Guo, Yu-Feng Li

Recent advances in LLMs, particularly in language reasoning and tool integration, have rapidly sparked the real-world development of Language Agents.

You Only Submit One Image to Find the Most Suitable Generative Model

no code implementations16 Dec 2024 Zhi Zhou, Lan-Zhe Guo, Peng-Xiao Song, Yu-Feng Li

In this paper, we propose a novel setting called Generative Model Identification (GMI), which aims to enable the user to identify the most appropriate generative model(s) for the user's requirements from a large number of candidate models efficiently.

Image Generation Text Matching

Fully Test-time Adaptation for Tabular Data

no code implementations14 Dec 2024 Zhi Zhou, Kun-Yang Yu, Lan-Zhe Guo, Yu-Feng Li

To this end, we propose the Fully Test-time Adaptation for Tabular data, namely FTAT, which enables FTTA methods to robustly optimize the label distribution of predictions, adapt to shifted covariate distributions, and suit a variety of tasks and models effectively.

Data Augmentation Test-time Adaptation

Enabling Small Models for Zero-Shot Selection and Reuse through Model Label Learning

no code implementations21 Aug 2024 Jia Zhang, Zhi Zhou, Lan-Zhe Guo, Yu-Feng Li

In this paper, we attempt to demonstrate that by constructing a model hub and aligning models with their functionalities using model labels, new tasks can be solved in a zero-shot manner by effectively selecting and reusing models in the hub.

Image Classification Zero-Shot Learning

Offline Imitation Learning with Model-based Reverse Augmentation

no code implementations18 Jun 2024 Jie-Jing Shao, Hao-Sen Shi, Lan-Zhe Guo, Yu-Feng Li

Specifically, we build a reverse dynamic model from the offline demonstrations, which can efficiently generate trajectories leading to the expert-observed states in a self-paced style.

Imitation Learning model

LawGPT: A Chinese Legal Knowledge-Enhanced Large Language Model

2 code implementations7 Jun 2024 Zhi Zhou, Jiang-Xin Shi, Peng-Xiao Song, Xiao-Wen Yang, Yi-Xuan Jin, Lan-Zhe Guo, Yu-Feng Li

Large language models (LLMs), including both proprietary and open-source models, have showcased remarkable capabilities in addressing a wide range of downstream tasks.

Language Modeling Language Modelling +1

Robustness Assessment of Mathematical Reasoning in the Presence of Missing and Contradictory Conditions

no code implementations7 Jun 2024 Shi-Yu Tian, Zhi Zhou, Lin-Han Jia, Lan-Zhe Guo, Yu-Feng Li

To further study this problem, we develop a benchmark called Problems with Missing and Contradictory conditions (PMC) and introduce two novel metrics to evaluate the performance of few-shot prompting methods in these scenarios.

Hallucination Mathematical Reasoning

DeCoOp: Robust Prompt Tuning with Out-of-Distribution Detection

1 code implementation1 Jun 2024 Zhi Zhou, Ming Yang, Jiang-Xin Shi, Lan-Zhe Guo, Yu-Feng Li

In this paper, we explore a problem setting called Open-world Prompt Tuning (OPT), which involves tuning prompts on base classes and evaluating on a combination of base and new classes.

Out-of-Distribution Detection

Investigating the Limitation of CLIP Models: The Worst-Performing Categories

no code implementations5 Oct 2023 Jie-Jing Shao, Jiang-Xin Shi, Xiao-Wen Yang, Lan-Zhe Guo, Yu-Feng Li

Contrastive Language-Image Pre-training (CLIP) provides a foundation model by integrating natural language into visual concepts, enabling zero-shot recognition on downstream tasks.

Prompt Engineering Zero-Shot Learning

LAMDA-SSL: Semi-Supervised Learning in Python

1 code implementation9 Aug 2022 Lin-Han Jia, Lan-Zhe Guo, Zhi Zhou, Yu-Feng Li

The second part shows the usage of LAMDA-SSL by abundant examples in detail.

Transfer and Share: Semi-Supervised Learning from Long-Tailed Data

no code implementations26 May 2022 Tong Wei, Qian-Yu Liu, Jiang-Xin Shi, Wei-Wei Tu, Lan-Zhe Guo

TRAS transforms the imbalanced pseudo-label distribution of a traditional SSL model via a delicate function to enhance the supervisory signals for minority classes.

Pseudo Label Representation Learning

Robust Deep Semi-Supervised Learning: A Brief Introduction

no code implementations12 Feb 2022 Lan-Zhe Guo, Zhi Zhou, Yu-Feng Li

Semi-supervised learning (SSL) is the branch of machine learning that aims to improve learning performance by leveraging unlabeled data when labels are insufficient.

STEP: Out-of-Distribution Detection in the Presence of Limited In-Distribution Labeled Data

no code implementations NeurIPS 2021 Zhi Zhou, Lan-Zhe Guo, Zhanzhan Cheng, Yu-Feng Li, ShiLiang Pu

However, in many real-world applications, it is desirable to have SSL algorithms that not only classify the samples drawn from the same distribution of labeled data but also detect out-of-distribution (OOD) samples drawn from an unknown distribution.

Out-of-Distribution Detection Out of Distribution (OOD) Detection

Weakly Supervised Learning Meets Ride-Sharing User Experience Enhancement

no code implementations19 Jan 2020 Lan-Zhe Guo, Feng Kuang, Zhang-Xun Liu, Yu-Feng Li, Nan Ma, Xiao-Hu Qie

For example, in user experience enhancement from Didi, one of the largest online ride-sharing platforms, the ride comment data contains severe label noise (due to the subjective factors of passengers) and severe label distribution bias (due to the sampling bias).

Weakly-supervised Learning

Reliable Weakly Supervised Learning: Maximize Gain and Maintain Safeness

no code implementations22 Apr 2019 Lan-Zhe Guo, Yu-Feng Li, Ming Li, Jin-Feng Yi, Bo-Wen Zhou, Zhi-Hua Zhou

We guide the optimization of label quality through a small amount of validation data, and to ensure the safeness of performance while maximizing performance gain.

Weakly-supervised Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.