Search Results for author: Yuxiao Dong

Found 73 papers, 55 papers with code

Extensive Self-Contrast Enables Feedback-Free Language Model Alignment

2 code implementations31 Mar 2024 Xiao Liu, Xixuan Song, Yuxiao Dong, Jie Tang

In this work, we introduce Self-Contrast, a feedback-free large language model alignment method via exploiting extensive self-generated negatives.

Language Modelling Large Language Model +1

Heterogeneous Graph Transformer

4 code implementations3 Mar 2020 Ziniu Hu, Yuxiao Dong, Kuansan Wang, Yizhou Sun

Recent years have witnessed the emerging success of graph neural networks (GNNs) for modeling structured data.

Graph Sampling Node Property Prediction

OGB-LSC: A Large-Scale Challenge for Machine Learning on Graphs

6 code implementations17 Mar 2021 Weihua Hu, Matthias Fey, Hongyu Ren, Maho Nakata, Yuxiao Dong, Jure Leskovec

Enabling effective and efficient machine learning (ML) over large-scale graph data (e. g., graphs with billions of edges) can have a great impact on both industrial and scientific applications.

BIG-bench Machine Learning Graph Learning +4

CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X

2 code implementations30 Mar 2023 Qinkai Zheng, Xiao Xia, Xu Zou, Yuxiao Dong, Shan Wang, Yufei Xue, Zihan Wang, Lei Shen, Andi Wang, Yang Li, Teng Su, Zhilin Yang, Jie Tang

Large pre-trained code generation models, such as OpenAI Codex, can generate syntax- and function-correct code, making the coding of programmers more productive and our pursuit of artificial general intelligence closer.

Code Generation

CogAgent: A Visual Language Model for GUI Agents

1 code implementation14 Dec 2023 Wenyi Hong, Weihan Wang, Qingsong Lv, Jiazheng Xu, Wenmeng Yu, Junhui Ji, Yan Wang, Zihan Wang, Yuxuan Zhang, Juanzi Li, Bin Xu, Yuxiao Dong, Ming Ding, Jie Tang

People are spending an enormous amount of time on digital devices through graphical user interfaces (GUIs), e. g., computer or smartphone screens.

Language Modelling Visual Question Answering

Network Embedding as Matrix Factorization: Unifying DeepWalk, LINE, PTE, and node2vec

4 code implementations9 Oct 2017 Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Kuansan Wang, Jie Tang

This work lays the theoretical foundation for skip-gram based network embedding methods, leading to a better understanding of latent network representation learning.

Network Embedding

Parameter-Efficient Prompt Tuning Makes Generalized and Calibrated Neural Text Retrievers

2 code implementations14 Jul 2022 Weng Lam Tam, Xiao Liu, Kaixuan Ji, Lilong Xue, Xingjian Zhang, Yuxiao Dong, Jiahua Liu, Maodi Hu, Jie Tang

By updating only 0. 1% of the model parameters, the prompt tuning strategy can help retrieval models achieve better generalization performance than traditional methods in which all parameters are updated.

Retrieval Text Retrieval +1

Open Graph Benchmark: Datasets for Machine Learning on Graphs

20 code implementations NeurIPS 2020 Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, Jure Leskovec

We present the Open Graph Benchmark (OGB), a diverse set of challenging and realistic benchmark datasets to facilitate scalable, robust, and reproducible graph machine learning (ML) research.

Knowledge Graphs Node Property Prediction

AgentBench: Evaluating LLMs as Agents

1 code implementation7 Aug 2023 Xiao Liu, Hao Yu, Hanchen Zhang, Yifan Xu, Xuanyu Lei, Hanyu Lai, Yu Gu, Hangliang Ding, Kaiwen Men, Kejuan Yang, Shudan Zhang, Xiang Deng, Aohan Zeng, Zhengxiao Du, Chenhui Zhang, Sheng Shen, Tianjun Zhang, Yu Su, Huan Sun, Minlie Huang, Yuxiao Dong, Jie Tang

We present AgentBench, a multi-dimensional evolving benchmark that currently consists of 8 distinct environments to assess LLM-as-Agent's reasoning and decision-making abilities in a multi-turn open-ended generation setting.

Decision Making Instruction Following

CogDL: A Comprehensive Library for Graph Deep Learning

1 code implementation1 Mar 2021 Yukuo Cen, Zhenyu Hou, Yan Wang, Qibin Chen, Yizhen Luo, Zhongming Yu, Hengrui Zhang, Xingcheng Yao, Aohan Zeng, Shiguang Guo, Yuxiao Dong, Yang Yang, Peng Zhang, Guohao Dai, Yu Wang, Chang Zhou, Hongxia Yang, Jie Tang

In CogDL, we propose a unified design for the training and evaluation of GNN models for various graph tasks, making it unique among existing graph learning libraries.

Graph Classification Graph Embedding +5

AgentTuning: Enabling Generalized Agent Abilities for LLMs

1 code implementation19 Oct 2023 Aohan Zeng, Mingdao Liu, Rui Lu, Bowen Wang, Xiao Liu, Yuxiao Dong, Jie Tang

Though many prompting methods have been proposed to complete particular agent tasks, there is lack of research focusing on improving the agent capabilities of LLMs themselves without compromising their general abilities.

Memorization

GPT-GNN: Generative Pre-Training of Graph Neural Networks

2 code implementations27 Jun 2020 Ziniu Hu, Yuxiao Dong, Kuansan Wang, Kai-Wei Chang, Yizhou Sun

Graph neural networks (GNNs) have been demonstrated to be powerful in modeling graph-structured data.

Attribute Graph Generation

LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding

1 code implementation28 Aug 2023 Yushi Bai, Xin Lv, Jiajie Zhang, Hongchang Lyu, Jiankai Tang, Zhidian Huang, Zhengxiao Du, Xiao Liu, Aohan Zeng, Lei Hou, Yuxiao Dong, Jie Tang, Juanzi Li

In this paper, we introduce LongBench, the first bilingual, multi-task benchmark for long context understanding, enabling a more rigorous evaluation of long context understanding.

16k Code Completion +2

GraphMAE: Self-Supervised Masked Graph Autoencoders

3 code implementations22 May 2022 Zhenyu Hou, Xiao Liu, Yukuo Cen, Yuxiao Dong, Hongxia Yang, Chunjie Wang, Jie Tang

Despite this, contrastive learning-which heavily relies on structural data augmentation and complicated training strategies-has been the dominant approach in graph SSL, while the progress of generative SSL on graphs, especially graph autoencoders (GAEs), has thus far not reached the potential as promised in other fields.

Contrastive Learning Graph Classification +4

GraphMAE2: A Decoding-Enhanced Masked Self-Supervised Graph Learner

2 code implementations10 Apr 2023 Zhenyu Hou, Yufei He, Yukuo Cen, Xiao Liu, Yuxiao Dong, Evgeny Kharlamov, Jie Tang

Graph self-supervised learning (SSL), including contrastive and generative approaches, offers great potential to address the fundamental challenge of label scarcity in real-world graph data.

Self-Supervised Learning

DeepInf: Social Influence Prediction with Deep Learning

1 code implementation15 Jul 2018 Jiezhong Qiu, Jian Tang, Hao Ma, Yuxiao Dong, Kuansan Wang, Jie Tang

Inspired by the recent success of deep neural networks in a wide range of computing applications, we design an end-to-end framework, DeepInf, to learn users' latent feature representation for predicting social influence.

Feature Engineering Representation Learning

Are we really making much progress? Revisiting, benchmarking, and refining heterogeneous graph neural networks

1 code implementation30 Dec 2021 Qingsong Lv, Ming Ding, Qiang Liu, Yuxiang Chen, Wenzheng Feng, Siming He, Chang Zhou, Jianguo Jiang, Yuxiao Dong, Jie Tang

Heterogeneous graph neural networks (HGNNs) have been blossoming in recent years, but the unique data processing and evaluation setups used by each work obstruct a full understanding of their advancements.

Benchmarking

Black-Box Prompt Optimization: Aligning Large Language Models without Model Training

1 code implementation7 Nov 2023 Jiale Cheng, Xiao Liu, Kehan Zheng, Pei Ke, Hongning Wang, Yuxiao Dong, Jie Tang, Minlie Huang

However, these models are often not well aligned with human intents, which calls for additional treatments on them, that is, the alignment problem.

CritiqueLLM: Scaling LLM-as-Critic for Effective and Explainable Evaluation of Large Language Model Generation

2 code implementations30 Nov 2023 Pei Ke, Bosi Wen, Zhuoer Feng, Xiao Liu, Xuanyu Lei, Jiale Cheng, Shengyuan Wang, Aohan Zeng, Yuxiao Dong, Hongning Wang, Jie Tang, Minlie Huang

Since the natural language processing (NLP) community started to make large language models (LLMs), such as GPT-4, act as a critic to evaluate the quality of generated texts, most of them only train a critique generation model of a specific scale on specific datasets.

Language Modelling Large Language Model

AutoWebGLM: Bootstrap And Reinforce A Large Language Model-based Web Navigating Agent

1 code implementation4 Apr 2024 Hanyu Lai, Xiao Liu, Iat Long Iong, Shuntian Yao, Yuxuan Chen, Pengbo Shen, Hao Yu, Hanchen Zhang, Xiaohan Zhang, Yuxiao Dong, Jie Tang

Large language models (LLMs) have fueled many intelligent agent tasks, such as web navigation -- but most existing agents perform far from satisfying in real-world webpages due to three factors: (1) the versatility of actions on webpages, (2) HTML text exceeding model processing capacity, and (3) the complexity of decision-making due to the open-domain nature of web.

Decision Making Language Modelling +1

MATCH: Metadata-Aware Text Classification in A Large Hierarchy

1 code implementation15 Feb 2021 Yu Zhang, Zhihong Shen, Yuxiao Dong, Kuansan Wang, Jiawei Han

Multi-label text classification refers to the problem of assigning each given document its most relevant labels from the label set.

General Classification Multi Label Text Classification +2

NetSMF: Large-Scale Network Embedding as Sparse Matrix Factorization

1 code implementation26 Jun 2019 Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Chi Wang, Kuansan Wang, Jie Tang

Previous research shows that 1) popular network embedding benchmarks, such as DeepWalk, are in essence implicitly factorizing a matrix with a closed form, and 2)the explicit factorization of such matrix generates more powerful embeddings than existing methods.

Network Embedding

CogCoM: Train Large Vision-Language Models Diving into Details through Chain of Manipulations

1 code implementation6 Feb 2024 Ji Qi, Ming Ding, Weihan Wang, Yushi Bai, Qingsong Lv, Wenyi Hong, Bin Xu, Lei Hou, Juanzi Li, Yuxiao Dong, Jie Tang

Vision-Language Models (VLMs) have demonstrated their widespread viability thanks to extensive training in aligning visual instructions to answers.

Visual Reasoning

LongAlign: A Recipe for Long Context Alignment of Large Language Models

1 code implementation31 Jan 2024 Yushi Bai, Xin Lv, Jiajie Zhang, Yuze He, Ji Qi, Lei Hou, Jie Tang, Yuxiao Dong, Juanzi Li

Extending large language models to effectively handle long contexts requires instruction fine-tuning on input sequences of similar length.

Instruction Following

Graph Robustness Benchmark: Benchmarking the Adversarial Robustness of Graph Machine Learning

1 code implementation8 Nov 2021 Qinkai Zheng, Xu Zou, Yuxiao Dong, Yukuo Cen, Da Yin, Jiarong Xu, Yang Yang, Jie Tang

To bridge this gap, we present the Graph Robustness Benchmark (GRB) with the goal of providing a scalable, unified, modular, and reproducible evaluation for the adversarial robustness of GML models.

Adversarial Robustness Benchmarking +1

Mask and Reason: Pre-Training Knowledge Graph Transformers for Complex Logical Queries

1 code implementation16 Aug 2022 Xiao Liu, Shiyu Zhao, Kai Su, Yukuo Cen, Jiezhong Qiu, Mengdi Zhang, Wei Wu, Yuxiao Dong, Jie Tang

In this work, we present the Knowledge Graph Transformer (kgTransformer) with masked pre-training and fine-tuning strategies.

OAG-BERT: Towards A Unified Backbone Language Model For Academic Knowledge Services

1 code implementation3 Mar 2021 Xiao Liu, Da Yin, Jingnan Zheng, Xingjian Zhang, Peng Zhang, Hongxia Yang, Yuxiao Dong, Jie Tang

Academic knowledge services have substantially facilitated the development of the science enterprise by providing a plenitude of efficient research tools.

Language Modelling Link Prediction

A Self-supervised Method for Entity Alignment

1 code implementation17 Jun 2021 Xiao Liu, Haoyun Hong, Xinghao Wang, Zeyi Chen, Evgeny Kharlamov, Yuxiao Dong, Jie Tang

We present SelfKG by leveraging this discovery to design a contrastive learning strategy across two KGs.

Contrastive Learning Entity Alignment +2

SelfKG: Self-Supervised Entity Alignment in Knowledge Graphs

1 code implementation2 Mar 2022 Xiao Liu, Haoyun Hong, Xinghao Wang, Zeyi Chen, Evgeny Kharlamov, Yuxiao Dong, Jie Tang

We present SelfKG with efficient strategies to optimize this objective for aligning entities without label supervision.

Entity Alignment Knowledge Graphs +1

SciGLM: Training Scientific Language Models with Self-Reflective Instruction Annotation and Tuning

1 code implementation15 Jan 2024 Dan Zhang, Ziniu Hu, Sining Zhoubian, Zhengxiao Du, Kaiyu Yang, Zihan Wang, Yisong Yue, Yuxiao Dong, Jie Tang

To bridge these gaps, we introduce SciGLM, a suite of scientific language models able to conduct college-level scientific reasoning.

Math Mathematical Reasoning

A Large-Scale Database for Graph Representation Learning

2 code implementations16 Nov 2020 Scott Freitas, Yuxiao Dong, Joshua Neil, Duen Horng Chau

With the rapid emergence of graph representation learning, the construction of new large-scale datasets is necessary to distinguish model capabilities and accurately assess the strengths and weaknesses of each technique.

Graph Representation Learning imbalanced classification

ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline

1 code implementation3 Apr 2024 Yifan Xu, Xiao Liu, Xinghan Liu, Zhenyu Hou, Yueyan Li, Xiaohan Zhang, Zihan Wang, Aohan Zeng, Zhengxiao Du, Wenyi Zhao, Jie Tang, Yuxiao Dong

Large language models (LLMs) have shown excellent mastering of human language, but still struggle in real-world applications that require mathematical problem-solving.

Math

Graph Contrastive Learning for Anomaly Detection

2 code implementations17 Aug 2021 Bo Chen, Jing Zhang, Xiaokang Zhang, Yuxiao Dong, Jian Song, Peng Zhang, Kaibo Xu, Evgeny Kharlamov, Jie Tang

To achieve the contrastive objective, we design a graph neural network encoder that can infer and further remove suspicious links during message passing, as well as learn the global context of the input graph.

Anomaly Detection Binary Classification +2

GRAND+: Scalable Graph Random Neural Networks

1 code implementation12 Mar 2022 Wenzheng Feng, Yuxiao Dong, Tinglin Huang, Ziqi Yin, Xu Cheng, Evgeny Kharlamov, Jie Tang

In this work, we present a scalable and high-performance GNN framework GRAND+ for semi-supervised graph learning.

Data Augmentation Graph Learning +2

Understanding WeChat User Preferences and "Wow" Diffusion

1 code implementation4 Mar 2021 Fanjin Zhang, Jie Tang, Xueyi Liu, Zhenyu Hou, Yuxiao Dong, Jing Zhang, Xiao Liu, Ruobing Xie, Kai Zhuang, Xu Zhang, Leyu Lin, Philip S. Yu

"Top Stories" is a novel friend-enhanced recommendation engine in WeChat, in which users can read articles based on preferences of both their own and their friends.

Graph Representation Learning Social and Information Networks

TDGIA:Effective Injection Attacks on Graph Neural Networks

1 code implementation12 Jun 2021 Xu Zou, Qinkai Zheng, Yuxiao Dong, Xinyu Guan, Evgeny Kharlamov, Jialiang Lu, Jie Tang

In the GIA scenario, the adversary is not able to modify the existing link structure and node attributes of the input graph, instead the attack is performed by injecting adversarial nodes into it.

Adversarial Attack

GOAL: A Challenging Knowledge-grounded Video Captioning Benchmark for Real-time Soccer Commentary Generation

1 code implementation26 Mar 2023 Ji Qi, Jifan Yu, Teng Tu, Kunyu Gao, Yifan Xu, Xinyu Guan, Xiaozhi Wang, Yuxiao Dong, Bin Xu, Lei Hou, Juanzi Li, Jie Tang, Weidong Guo, Hui Liu, Yu Xu

Despite the recent emergence of video captioning models, how to generate vivid, fine-grained video descriptions based on the background knowledge (i. e., long and informative commentary about the domain-specific scenes with appropriate reasoning) is still far from being solved, which however has great applications such as automatic sports narrative.

Video Captioning

Open-World Semi-Supervised Learning for Node Classification

1 code implementation18 Mar 2024 Yanling Wang, Jing Zhang, Lingxi Zhang, Lixin Liu, Yuxiao Dong, Cuiping Li, Hong Chen, Hongzhi Yin

Open-world semi-supervised learning (Open-world SSL) for node classification, that classifies unlabeled nodes into seen classes or multiple novel classes, is a practical but under-explored problem in the graph community.

Classification Contrastive Learning +2

Will This Paper Increase Your h-index? Scientific Impact Prediction

2 code implementations15 Dec 2014 Yuxiao Dong, Reid A. Johnson, Nitesh V. Chawla

The effectiveness of such predictions, however, is fundamentally limited by the power-law distribution of citations, whereby publications with few citations are extremely common and publications with many citations are relatively rare.

Social and Information Networks Digital Libraries Physics and Society H.2.8; H.3.7

Neural Tensor Factorization

no code implementations13 Feb 2018 Xian Wu, Baoxu Shi, Yuxiao Dong, Chao Huang, Nitesh Chawla

Neural collaborative filtering (NCF) and recurrent recommender systems (RRN) have been successful in modeling user-item relational data.

Collaborative Filtering Link Prediction +1

Revisiting Knowledge Base Embedding as Tensor Decomposition

no code implementations ICLR 2018 Jiezhong Qiu, Hao Ma, Yuxiao Dong, Kuansan Wang, Jie Tang

We study the problem of knowledge base (KB) embedding, which is usually addressed through two frameworks---neural KB embedding and tensor decomposition.

Link Prediction Tensor Decomposition

A Century of Science: Globalization of Scientific Collaborations, Citations, and Innovations

no code implementations17 Apr 2017 Yuxiao Dong, Hao Ma, Zhihong Shen, Kuansan Wang

We find that science has benefited from the shift from individual work to collaborative effort, with over 90% of the world-leading innovations generated by collaborations in this century, nearly four times higher than they were in the 1900s.

Digital Libraries Social and Information Networks Physics and Society

Generalizing Graph Convolutional Networks

1 code implementation1 Jan 2021 Jialin Zhao, Yuxiao Dong, Jie Tang, Ming Ding, Kuansan Wang

Graph convolutional networks (GCNs) have emerged as a powerful framework for mining and learning with graphs.

P-INT: A Path-based Interaction Model for Few-shot Knowledge Graph Completion

no code implementations Findings (EMNLP) 2021 Jingwen Xu, Jing Zhang, Xirui Ke, Yuxiao Dong, Hong Chen, Cuiping Li, Yongbin Liu

Its general process is to first encode the implicit relation of an entity pair and then match the relation of a query entity pair with the relations of the reference entity pairs.

Knowledge Graph Completion Relation

Adaptive Diffusion in Graph Neural Networks

no code implementations NeurIPS 2021 Jialin Zhao, Yuxiao Dong, Ming Ding, Evgeny Kharlamov, Jie Tang

Notably, message passing based GNNs, e. g., graph convolutional networks, leverage the immediate neighbors of each node during the aggregation process, and recently, graph diffusion convolution (GDC) is proposed to expand the propagation neighborhood by leveraging generalized graph diffusion.

Diagonal Graph Convolutional Networks with Adaptive Neighborhood Aggregation

no code implementations25 Sep 2019 Jie Zhang, Yuxiao Dong, Jie Tang

In this paper, we revisit the mathematical foundation of GCNs and study how to extend their representation capacity.

Graph Attention Graph Classification +1

IDPG: An Instance-Dependent Prompt Generation Method

no code implementations NAACL 2022 Zhuofeng Wu, Sinong Wang, Jiatao Gu, Rui Hou, Yuxiao Dong, V. G. Vinod Vydiswaran, Hao Ma

Prompt tuning is a new, efficient NLP transfer learning paradigm that adds a task-specific prompt in each input instance during the model training stage.

Language Modelling Natural Language Understanding +2

Revisiting Parallel Context Windows: A Frustratingly Simple Alternative and Chain-of-Thought Deterioration

no code implementations24 May 2023 Kejuan Yang, Xiao Liu, Kaiwen Men, Aohan Zeng, Yuxiao Dong, Jie Tang

We identify two crucial limitations in the evaluation of recent parallel-integrated method Parallel Context Windows (PCW), which extends the maximum context lengths of language models, e. g., 2048 for LLaMA, by harnessing window-wise attention and positional embedding techniques.

Long-Context Understanding

APAR: LLMs Can Do Auto-Parallel Auto-Regressive Decoding

no code implementations12 Jan 2024 Mingdao Liu, Aohan Zeng, Bowen Wang, Peng Zhang, Jie Tang, Yuxiao Dong

The massive adoption of large language models (LLMs) demands efficient deployment strategies.

xTrimoPGLM: Unified 100B-Scale Pre-trained Transformer for Deciphering the Language of Protein

no code implementations11 Jan 2024 Bo Chen, Xingyi Cheng, Pan Li, Yangli-ao Geng, Jing Gong, Shen Li, Zhilei Bei, Xu Tan, Boyan Wang, Xin Zeng, Chiming Liu, Aohan Zeng, Yuxiao Dong, Jie Tang, Le Song

We propose a unified protein language model, xTrimoPGLM, to address these two types of tasks simultaneously through an innovative pre-training framework.

Protein Language Model

TriSampler: A Better Negative Sampling Principle for Dense Retrieval

no code implementations19 Feb 2024 Zhen Yang, Zhou Shao, Yuxiao Dong, Jie Tang

Negative sampling stands as a pivotal technique in dense retrieval, essential for training effective retrieval models and significantly impacting retrieval performance.

Retrieval

Middleware for LLMs: Tools Are Instrumental for Language Agents in Complex Environments

no code implementations22 Feb 2024 Yu Gu, Yiheng Shu, Hao Yu, Xiao Liu, Yuxiao Dong, Jie Tang, Jayanth Srinivasa, Hugo Latapie, Yu Su

The applications of large language models (LLMs) have expanded well beyond the confines of text processing, signaling a new era where LLMs are envisioned as generalist language agents capable of operating within complex real-world environments.

OAG-Bench: A Human-Curated Benchmark for Academic Graph Mining

no code implementations24 Feb 2024 Fanjin Zhang, Shijie Shi, Yifan Zhu, Bo Chen, Yukuo Cen, Jifan Yu, Yelin Chen, Lulu Wang, Qingfei Zhao, Yuqing Cheng, Tianyi Han, Yuwei An, Dan Zhang, Weng Lam Tam, Kun Cao, Yunhe Pang, Xinyu Guan, Huihui Yuan, Jian Song, Xiaoyan Li, Yuxiao Dong, Jie Tang

We envisage that OAG-Bench can serve as a common ground for the community to evaluate and compare algorithms in academic graph mining, thereby accelerating algorithm development and advancement in this field.

Graph Mining

Understanding Emergent Abilities of Language Models from the Loss Perspective

no code implementations23 Mar 2024 Zhengxiao Du, Aohan Zeng, Yuxiao Dong, Jie Tang

Recent studies have put into question the belief that emergent abilities in language models are exclusive to large models.

ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback

no code implementations1 Apr 2024 Zhenyu Hou, Yilin Niu, Zhengxiao Du, Xiaohan Zhang, Xiao Liu, Aohan Zeng, Qinkai Zheng, Minlie Huang, Hongning Wang, Jie Tang, Yuxiao Dong

The work presents our practices of aligning LLMs with human preferences, offering insights into the challenges and solutions in RLHF implementations.

Cannot find the paper you are looking for? You can Submit a new open access paper.