Search Results for author: Haotian Ye

Found 18 papers, 7 papers with code

DOF: Accelerating High-order Differential Operators with Forward Propagation

no code implementations15 Feb 2024 Ruichen Li, Chuwei Wang, Haotian Ye, Di He, LiWei Wang

Solving partial differential equations (PDEs) efficiently is essential for analyzing complex physical systems.

Selecting Large Language Model to Fine-tune via Rectified Scaling Law

no code implementations4 Feb 2024 Haowei Lin, Baizhou Huang, Haotian Ye, Qinyu Chen, ZiHao Wang, Sujian Li, Jianzhu Ma, Xiaojun Wan, James Zou, Yitao Liang

The ever-growing ecosystem of LLMs has posed a challenge in selecting the most appropriate pre-trained model to fine-tune amidst a sea of options.

Language Modelling Large Language Model

TransliCo: A Contrastive Learning Framework to Address the Script Barrier in Multilingual Pretrained Language Models

1 code implementation12 Jan 2024 Yihong Liu, Chunlan Ma, Haotian Ye, Hinrich Schütze

As a result, mPLMs present a script barrier: representations from different scripts are located in different subspaces, which is a strong indicator of why crosslingual transfer involving languages of different scripts shows sub-optimal performance.

Contrastive Learning Transliteration

MoSECroT: Model Stitching with Static Word Embeddings for Crosslingual Zero-shot Transfer

no code implementations9 Jan 2024 Haotian Ye, Yihong Liu, Chunlan Ma, Hinrich Schütze

In this paper, we introduce MoSECroT Model Stitching with Static Word Embeddings for Crosslingual Zero-shot Transfer), a novel and challenging task that is especially relevant to low-resource languages for which static word embeddings are available.

Word Embeddings

In-context Vectors: Making In Context Learning More Effective and Controllable Through Latent Space Steering

1 code implementation11 Nov 2023 Sheng Liu, Haotian Ye, Lei Xing, James Zou

On a new query, instead of adding demonstrations to the prompt, we shift the latent states of the LLM using the ICV.

In-Context Learning Style Transfer

Towards Revealing the Mystery behind Chain of Thought: A Theoretical Perspective

no code implementations NeurIPS 2023 Guhao Feng, Bohang Zhang, Yuntian Gu, Haotian Ye, Di He, LiWei Wang

By using circuit complexity theory, we first give impossibility results showing that bounded-depth Transformers are unable to directly produce correct answers for basic arithmetic/equation tasks unless the model size grows super-polynomially with respect to the input length.

Decision Making Math

A study of conceptual language similarity: comparison and evaluation

no code implementations22 May 2023 Haotian Ye, Yihong Liu, Hinrich Schütze

An interesting line of research in natural language processing (NLP) aims to incorporate linguistic typology to bridge linguistic diversity and assist the research of low-resource languages.

Binary Classification

A Crosslingual Investigation of Conceptualization in 1335 Languages

3 code implementations15 May 2023 Yihong Liu, Haotian Ye, Leonie Weissweiler, Philipp Wicke, Renhao Pei, Robert Zangenfeind, Hinrich Schütze

The resulting measure for the conceptual similarity of two languages is complementary to standard genealogical, typological, and surface similarity measures.

Taxi1500: A Multilingual Dataset for Text Classification in 1500 Languages

no code implementations15 May 2023 Chunlan Ma, Ayyoob ImaniGooghari, Haotian Ye, Ehsaneddin Asgari, Hinrich Schütze

While natural language processing tools have been developed extensively for some of the world's languages, a significant portion of the world's over 7000 languages are still neglected.

text-classification Text Classification

Discovering Latent Knowledge in Language Models Without Supervision

1 code implementation7 Dec 2022 Collin Burns, Haotian Ye, Dan Klein, Jacob Steinhardt

Existing techniques for training language models can be misaligned with the truth: if we train models with imitation learning, they may reproduce errors that humans make; if we train them to generate text that humans rate highly, they may output errors that human evaluators can't detect.

Imitation Learning Language Modelling +2

Freeze then Train: Towards Provable Representation Learning under Spurious Correlations and Feature Noise

1 code implementation20 Oct 2022 Haotian Ye, James Zou, Linjun Zhang

This opens a promising strategy to first train a feature learner rather than a classifier, and then perform linear probing (last layer retraining) in the test environment.

Representation Learning

On the Power of Pre-training for Generalization in RL: Provable Benefits and Hardness

no code implementations19 Oct 2022 Haotian Ye, Xiaoyu Chen, LiWei Wang, Simon S. Du

Generalization in Reinforcement Learning (RL) aims to learn an agent during training that generalizes to the target environment.

Reinforcement Learning (RL)

Towards a Theoretical Framework of Out-of-Distribution Generalization

no code implementations NeurIPS 2021 Haotian Ye, Chuanlong Xie, Tianle Cai, Ruichen Li, Zhenguo Li, LiWei Wang

We also introduce a new concept of expansion function, which characterizes to what extent the variance is amplified in the test domains over the training domains, and therefore give a quantitative meaning of invariant features.

Domain Generalization Model Selection +1

Risk Variance Penalization

no code implementations13 Jun 2020 Chuanlong Xie, Haotian Ye, Fei Chen, Yue Liu, Rui Sun, Zhenguo Li

The key of the out-of-distribution (OOD) generalization is to generalize invariance from training domains to target domains.

Cannot find the paper you are looking for? You can Submit a new open access paper.