Search Results for author: Haotian Ye

Found 27 papers, 15 papers with code

A Foundational Generative Model for Breast Ultrasound Image Analysis

no code implementations12 Jan 2025 Haojun Yu, Youcheng Li, Nan Zhang, Zihan Niu, Xuantong Gong, Yanwen Luo, Haotian Ye, Siyu He, Quanlin Wu, Wangyan Qin, Mengyuan Zhou, Jie Han, Jia Tao, Ziwei Zhao, Di Dai, Di He, Dong Wang, Binghui Tang, Ling Huo, James Zou, Qingli Zhu, Yong Wang, LiWei Wang

Additionally, we characterized the scaling effect of using generated data which was as effective as the collected real-world data for training diagnostic models.

A Federated Approach to Few-Shot Hate Speech Detection for Marginalized Communities

no code implementations6 Dec 2024 Haotian Ye, Axel Wisiorek, Antonis Maronikolakis, Özge Alaçam, Hinrich Schütze

In this paper, we aim to provide marginalized communities living in societies where the dominant language is low-resource with a privacy-preserving tool to protect themselves from hate speech on the internet by filtering offensive content in their native languages.

Federated Learning Few-Shot Learning +2

Reducing Hallucinations in Vision-Language Models via Latent Space Steering

1 code implementation21 Oct 2024 Sheng Liu, Haotian Ye, Lei Xing, James Zou

Unlike in large language models (LLMs), hallucination in LVLMs often arises from misalignments between visual inputs and textual outputs.

Hallucination

Geometric Trajectory Diffusion Models

1 code implementation16 Oct 2024 Jiaqi Han, Minkai Xu, Aaron Lou, Haotian Ye, Stefano Ermon

In this work, we propose geometric trajectory diffusion models (GeoTDM), the first diffusion model for modeling the temporal distribution of 3D geometric trajectories.

Protein Design

LangSAMP: Language-Script Aware Multilingual Pretraining

1 code implementation26 Sep 2024 Yihong Liu, Haotian Ye, Chunlan Ma, Mingyang Wang, Hinrich Schütze

However, this removal increases the burden on token embeddings to encode all language-specific information, which may hinder the model's ability to produce more language-neutral representations.

Continual Pretraining Language Modeling +2

How Transliterations Improve Crosslingual Alignment

1 code implementation25 Sep 2024 Yihong Liu, Mingyang Wang, Amir Hossein Kargaran, Ayyoob Imani, Orgest Xhelili, Haotian Ye, Chunlan Ma, François Yvon, Hinrich Schütze

However, we also show that better alignment does not always yield better downstream performance, suggesting that further research is needed to clarify the connection between alignment and performance.

Sentence Transliteration

TFG: Unified Training-Free Guidance for Diffusion Models

1 code implementation24 Sep 2024 Haotian Ye, Haowei Lin, Jiaqi Han, Minkai Xu, Sheng Liu, Yitao Liang, Jianzhu Ma, James Zou, Stefano Ermon

Given an unconditional diffusion model and a predictor for a target property of interest (e. g., a classifier), the goal of training-free guidance is to generate samples with desirable target properties without additional training.

Exploring the Role of Transliteration in In-Context Learning for Low-resource Languages Written in Non-Latin Scripts

no code implementations2 Jul 2024 Chunlan Ma, Yihong Liu, Haotian Ye, Hinrich Schütze

Inspired by recent work that leverages transliteration in encoder-only models, we investigate whether transliteration is also effective in improving LLMs' performance for low-resource languages written in non-Latin scripts.

Decoder In-Context Learning +3

TransMI: A Framework to Create Strong Baselines from Multilingual Pretrained Language Models for Transliterated Data

1 code implementation16 May 2024 Yihong Liu, Chunlan Ma, Haotian Ye, Hinrich Schütze

TransMI can create strong baselines for data that is transliterated into a common script by exploiting an existing mPLM and its tokenizer without any training.

Transliteration

DOF: Accelerating High-order Differential Operators with Forward Propagation

no code implementations15 Feb 2024 Ruichen Li, Chuwei Wang, Haotian Ye, Di He, LiWei Wang

Solving partial differential equations (PDEs) efficiently is essential for analyzing complex physical systems.

Selecting Large Language Model to Fine-tune via Rectified Scaling Law

no code implementations4 Feb 2024 Haowei Lin, Baizhou Huang, Haotian Ye, Qinyu Chen, ZiHao Wang, Sujian Li, Jianzhu Ma, Xiaojun Wan, James Zou, Yitao Liang

The ever-growing ecosystem of LLMs has posed a challenge in selecting the most appropriate pre-trained model to fine-tune amidst a sea of options.

Language Modeling Language Modelling +1

TransliCo: A Contrastive Learning Framework to Address the Script Barrier in Multilingual Pretrained Language Models

1 code implementation12 Jan 2024 Yihong Liu, Chunlan Ma, Haotian Ye, Hinrich Schütze

As a consequence, mPLMs are faced with a script barrier: representations from different scripts are located in different subspaces, which can result in crosslingual transfer involving languages of different scripts performing suboptimally.

Contrastive Learning Transliteration

MoSECroT: Model Stitching with Static Word Embeddings for Crosslingual Zero-shot Transfer

no code implementations9 Jan 2024 Haotian Ye, Yihong Liu, Chunlan Ma, Hinrich Schütze

In this paper, we introduce MoSECroT Model Stitching with Static Word Embeddings for Crosslingual Zero-shot Transfer), a novel and challenging task that is especially relevant to low-resource languages for which static word embeddings are available.

Word Embeddings

In-context Vectors: Making In Context Learning More Effective and Controllable Through Latent Space Steering

1 code implementation11 Nov 2023 Sheng Liu, Haotian Ye, Lei Xing, James Zou

On a new query, instead of adding demonstrations to the prompt, we shift the latent states of the LLM using the ICV.

In-Context Learning Style Transfer

Towards Revealing the Mystery behind Chain of Thought: A Theoretical Perspective

no code implementations NeurIPS 2023 Guhao Feng, Bohang Zhang, Yuntian Gu, Haotian Ye, Di He, LiWei Wang

By using circuit complexity theory, we first give impossibility results showing that bounded-depth Transformers are unable to directly produce correct answers for basic arithmetic/equation tasks unless the model size grows super-polynomially with respect to the input length.

Decision Making Math

A study of conceptual language similarity: comparison and evaluation

no code implementations22 May 2023 Haotian Ye, Yihong Liu, Hinrich Schütze

An interesting line of research in natural language processing (NLP) aims to incorporate linguistic typology to bridge linguistic diversity and assist the research of low-resource languages.

Binary Classification Diversity

A Crosslingual Investigation of Conceptualization in 1335 Languages

3 code implementations15 May 2023 Yihong Liu, Haotian Ye, Leonie Weissweiler, Philipp Wicke, Renhao Pei, Robert Zangenfeind, Hinrich Schütze

The resulting measure for the conceptual similarity of two languages is complementary to standard genealogical, typological, and surface similarity measures.

Taxi1500: A Multilingual Dataset for Text Classification in 1500 Languages

1 code implementation15 May 2023 Chunlan Ma, Ayyoob ImaniGooghari, Haotian Ye, Renhao Pei, Ehsaneddin Asgari, Hinrich Schütze

While natural language processing tools have been developed extensively for some of the world's languages, a significant portion of the world's over 7000 languages are still neglected.

text-classification Text Classification

Discovering Latent Knowledge in Language Models Without Supervision

1 code implementation7 Dec 2022 Collin Burns, Haotian Ye, Dan Klein, Jacob Steinhardt

Existing techniques for training language models can be misaligned with the truth: if we train models with imitation learning, they may reproduce errors that humans make; if we train them to generate text that humans rate highly, they may output errors that human evaluators can't detect.

Imitation Learning Language Modelling +2

Freeze then Train: Towards Provable Representation Learning under Spurious Correlations and Feature Noise

1 code implementation20 Oct 2022 Haotian Ye, James Zou, Linjun Zhang

This opens a promising strategy to first train a feature learner rather than a classifier, and then perform linear probing (last layer retraining) in the test environment.

Representation Learning

On the Power of Pre-training for Generalization in RL: Provable Benefits and Hardness

no code implementations19 Oct 2022 Haotian Ye, Xiaoyu Chen, LiWei Wang, Simon S. Du

Generalization in Reinforcement Learning (RL) aims to learn an agent during training that generalizes to the target environment.

Reinforcement Learning (RL)

Towards a Theoretical Framework of Out-of-Distribution Generalization

no code implementations NeurIPS 2021 Haotian Ye, Chuanlong Xie, Tianle Cai, Ruichen Li, Zhenguo Li, LiWei Wang

We also introduce a new concept of expansion function, which characterizes to what extent the variance is amplified in the test domains over the training domains, and therefore give a quantitative meaning of invariant features.

Domain Generalization Model Selection +1

Risk Variance Penalization

no code implementations13 Jun 2020 Chuanlong Xie, Haotian Ye, Fei Chen, Yue Liu, Rui Sun, Zhenguo Li

The key of the out-of-distribution (OOD) generalization is to generalize invariance from training domains to target domains.

Cannot find the paper you are looking for? You can Submit a new open access paper.