Search Results for author: Boshi Wang

Found 17 papers, 11 papers with code

Is the Reversal Curse a Binding Problem? Uncovering Limitations of Transformers from a Basic Generalization Failure

1 code implementation2 Apr 2025 Boshi Wang, Huan Sun

Despite their impressive capabilities, LLMs exhibit a basic generalization failure known as the Reversal Curse, where they struggle to learn reversible factual associations.

Arithmetic Reasoning Data Augmentation

Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization

1 code implementation23 May 2024 Boshi Wang, Xiang Yue, Yu Su, Huan Sun

The levels of generalization also vary across reasoning types: when faced with out-of-distribution examples, transformers fail to systematically generalize for composition but succeed for comparison.

Beamforming Design for Double-Active-RIS-aided Communication Systems with Inter-Excitation

no code implementations17 Mar 2024 Boshi Wang, Cunhua Pan, Hong Ren, Zhiyuan Yu, Yang Zhang, Mengyu Liu, Gui Zhou

Furthermore, the results demonstrate that the proposed scheme can enhance the WSR by 30\% compared to scenarios that do not take this effect into account when the maximum amplification gain is 40 dB.

LLMs in the Imaginarium: Tool Learning through Simulated Trial and Error

1 code implementation7 Mar 2024 Boshi Wang, Hao Fang, Jason Eisner, Benjamin Van Durme, Yu Su

We find that existing LLMs, including GPT-4 and open-source LLMs specifically fine-tuned for tool use, only reach a correctness rate in the range of 30% to 60%, far from reliable use in practice.

Continual Learning In-Context Learning

Joint Beamforming Design for Double Active RIS-assisted Radar-Communication Coexistence Systems

no code implementations7 Feb 2024 Mengyu Liu, Hong Ren, Cunhua Pan, Boshi Wang, Zhiyuan Yu, Ruisong Weng, Kangda Zhi, Yongchao He

However, when radar and communication equipment coexist in the same system, i. e. radar-communication coexistence (RCC), the interference from communication systems to radar can be large and cannot be ignored.

Integrated sensing and communication ISAC

Secure Wireless Communication in Active RIS-Assisted DFRC System

no code implementations3 Feb 2024 Yang Zhang, Hong Ren, Cunhua Pan, Boshi Wang, Zhiyuan Yu, Ruisong Weng, Tuo Wu, Yongchao He

This work considers a dual-functional radar and communication (DFRC) system with an active reconfigurable intelligent surface (RIS) and a potential eavesdropper.

How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their Vulnerabilities

1 code implementation15 Nov 2023 Lingbo Mo, Boshi Wang, Muhao Chen, Huan Sun

The rapid progress in open-source Large Language Models (LLMs) is significantly driving AI development forward.

Ethics Fairness +3

Mind2Web: Towards a Generalist Agent for the Web

1 code implementation NeurIPS 2023 Xiang Deng, Yu Gu, Boyuan Zheng, Shijie Chen, Samuel Stevens, Boshi Wang, Huan Sun, Yu Su

We introduce Mind2Web, the first dataset for developing and evaluating generalist agents for the web that can follow language instructions to complete complex tasks on any website.

Can ChatGPT Defend its Belief in Truth? Evaluating LLM Reasoning via Debate

no code implementations22 May 2023 Boshi Wang, Xiang Yue, Huan Sun

Large language models (LLMs) such as ChatGPT and GPT-4 have shown impressive performance in complex reasoning tasks.

Benchmarking Math +1

Automatic Evaluation of Attribution by Large Language Models

1 code implementation10 May 2023 Xiang Yue, Boshi Wang, Ziru Chen, Kai Zhang, Yu Su, Huan Sun

We manually curate a set of test examples covering 12 domains from a generative search engine, New Bing.

Fact Checking Language Modeling +4

Scalable Learning with Incremental Probabilistic PCA

1 code implementation IEEE International Conference on Big Data 2022 Boshi Wang, Adrian Barbu

Incremental class learning is the classification problem of learning a model where instances from new object classes are added sequentially, and it is desired that the model be retrained only on the new classes with minimal training on the old classes.

class-incremental learning Class Incremental Learning +1

Towards Understanding Chain-of-Thought Prompting: An Empirical Study of What Matters

2 code implementations20 Dec 2022 Boshi Wang, Sewon Min, Xiang Deng, Jiaming Shen, You Wu, Luke Zettlemoyer, Huan Sun

Chain-of-Thought (CoT) prompting can dramatically improve the multi-step reasoning abilities of large language models (LLMs).

A Retrieve-and-Read Framework for Knowledge Graph Link Prediction

1 code implementation19 Dec 2022 Vardaan Pahuja, Boshi Wang, Hugo Latapie, Jayanth Srinivasa, Yu Su

To address the limitations of existing KG link prediction frameworks, we propose a novel retrieve-and-read framework, which first retrieves a relevant subgraph context for the query and then jointly reasons over the context and the query with a high-capacity reader.

Knowledge Graph Completion Link Prediction +1

Iteratively Prompt Pre-trained Language Models for Chain of Thought

1 code implementation16 Mar 2022 Boshi Wang, Xiang Deng, Huan Sun

While Pre-trained Language Models (PLMs) internalize a great amount of world knowledge, they have been shown incapable of recalling these knowledge to solve tasks requiring complex & multi-step reasoning.

World Knowledge

Automatic Loss Function Search for Predict-Then-Optimize Problems with Strong Ranking Property

no code implementations ICLR 2022 Boshi Wang, Jialin Yi, Hang Dong, Bo Qiao, Chuan Luo, QIngwei Lin

Combinatorial optimization problems with parameters to be predicted from side information are commonly seen in a variety of problems during the paradigm shift from reactive decision making to proactive decision making.

Combinatorial Optimization Decision Making +1

Cannot find the paper you are looking for? You can Submit a new open access paper.