Search Results for author: ZiYi Yang

Found 68 papers, 26 papers with code

UniGeo: Taming Video Diffusion for Unified Consistent Geometry Estimation

no code implementations30 May 2025 Yang-tian Sun, Xin Yu, Zehuan Huang, Yi-Hua Huang, Yuan-Chen Guo, ZiYi Yang, Yan-Pei Cao, Xiaojuan Qi

Recently, methods leveraging diffusion model priors to assist monocular geometric estimation (e. g., depth and normal) have gained significant attention due to their strong generalization ability.

Video Generation

QwenLong-L1: Towards Long-Context Large Reasoning Models with Reinforcement Learning

1 code implementation23 May 2025 Fanqi Wan, Weizhou Shen, Shengyi Liao, Yingcheng Shi, Chenliang Li, ZiYi Yang, Ji Zhang, Fei Huang, Jingren Zhou, Ming Yan

To bridge this gap, we first formalize the paradigm of long-context reasoning RL, and identify key challenges in suboptimal training efficiency and unstable optimization process.

Question Answering Reinforcement Learning (RL)

ThinkSwitcher: When to Think Hard, When to Think Fast

no code implementations20 May 2025 Guosheng Liang, Longguang Zhong, ZiYi Yang, Xiaojun Quan

To leverage this capability, we propose ThinkSwitcher, a framework that enables a single LRM to dynamically switch between short and long CoT modes based on task complexity.

Mutual-Taught for Co-adapting Policy and Reward Models

no code implementations17 May 2025 Tianyuan Shi, Canbin Huang, Fanqi Wan, Longguang Zhong, ZiYi Yang, Weizhou Shen, Xiaojun Quan, Ming Yan

During the preference optimization of large language models (LLMs), distribution shifts may arise between newly generated model samples and the data used to train the reward model (RM).

FuseRL: Dense Preference Optimization for Heterogeneous Model Fusion

no code implementations9 Apr 2025 Longguang Zhong, Fanqi Wan, ZiYi Yang, Guosheng Liang, Tianyuan Shi, Xiaojun Quan

Heterogeneous model fusion enhances the performance of LLMs by integrating the knowledge and capabilities of multiple structurally diverse models.

Scaling Laws of Synthetic Data for Language Models

no code implementations25 Mar 2025 Zeyu Qin, Qingxiu Dong, Xingxing Zhang, Li Dong, Xiaolong Huang, ZiYi Yang, Mahmoud Khademi, Dongdong Zhang, Hany Hassan Awadalla, Yi R. Fung, Weizhu Chen, Minhao Cheng, Furu Wei

Key findings from our extensive mathematical experiments on SynthLLM include: (1) SynthLLM generates synthetic data that reliably adheres to the rectified scaling law across various model sizes; (2) Performance improvements plateau near 300B tokens; and (3) Larger models approach optimal performance with fewer training tokens.

Synthetic Data Generation

FuseChat-3.0: Preference Optimization Meets Heterogeneous Model Fusion

1 code implementation6 Mar 2025 ZiYi Yang, Fanqi Wan, Longguang Zhong, Canbin Huang, Guosheng Liang, Xiaojun Quan

The FuseChat-3. 0 training pipeline consists of two key stages: (1) supervised fine-tuning (SFT) to align the target and source model distributions, and (2) Direct Preference Optimization (DPO) to apply preferences from multiple source LLMs to fine-tune the target model.

General Knowledge Instruction Following +1

Frequency-aware Event Cloud Network

no code implementations30 Dec 2024 Hongwei Ren, Fei Ma, Xiaopeng Lin, Yuetong Fang, Hongxiang Huang, Yulong Huang, Yue Zhou, Haotian Fu, ZiYi Yang, Fei Richard Yu, Bojun Cheng

Event cameras are biologically inspired sensors that emit events asynchronously with remarkable temporal resolution, garnering significant attention from both industry and academia.

Action Recognition Pose Estimation

Deformable Radial Kernel Splatting

no code implementations CVPR 2025 Yi-Hua Huang, Ming-Xian Lin, Yang-tian Sun, ZiYi Yang, Xiaoyang Lyu, Yan-Pei Cao, Xiaojuan Qi

Recently, Gaussian splatting has emerged as a robust technique for representing 3D scenes, enabling real-time rasterization and high-fidelity rendering.

Weighted-Reward Preference Optimization for Implicit Model Fusion

4 code implementations4 Dec 2024 ZiYi Yang, Fanqi Wan, Longguang Zhong, Tianyuan Shi, Xiaojun Quan

To address distributional deviations between the source and target LLMs, WRPO introduces a progressive adaptation strategy that gradually shifts reliance on preferred examples from the target LLM to the source LLMs.

model

OASIS: Open Agent Social Interaction Simulations with One Million Agents

1 code implementation18 Nov 2024 ZiYi Yang, Zaibin Zhang, Zirui Zheng, Yuxian Jiang, Ziyue Gan, Zhiyu Wang, Zijian Ling, Jinsong Chen, Martz Ma, Bowen Dong, Prateek Gupta, Shuyue Hu, Zhenfei Yin, Guohao Li, Xu Jia, Lijun Wang, Bernard Ghanem, Huchuan Lu, Chaochao Lu, Wanli Ouyang, Yu Qiao, Philip Torr, Jing Shao

There has been a growing interest in enhancing rule-based agent-based models (ABMs) for social media platforms (i. e., X, Reddit) with more realistic large language model (LLM) agents, thereby allowing for a more nuanced study of complex systems.

Large Language Model Recommendation Systems

Towards Realistic Example-based Modeling via 3D Gaussian Stitching

no code implementations CVPR 2025 Xinyu Gao, ZiYi Yang, Bingchen Gong, Xiaoguang Han, Sipeng Yang, Xiaogang Jin

To this end, we present an example-based modeling method that combines multiple Gaussian fields in a point-based representation using sample-guided synthesis.

3DGS

See What LLMs Cannot Answer: A Self-Challenge Framework for Uncovering LLM Weaknesses

1 code implementation16 Aug 2024 Yulong Chen, Yang Liu, Jianhao Yan, Xuefeng Bai, Ming Zhong, Yinghao Yang, ZiYi Yang, Chenguang Zhu, Yue Zhang

We then build a benchmark, SC-G4, consisting of 1, 835 instances generated by GPT-4 using these patterns, with human-annotated gold responses.

FuseChat: Knowledge Fusion of Chat Models

3 code implementations15 Aug 2024 Fanqi Wan, Longguang Zhong, ZiYi Yang, Ruijun Chen, Xiaojun Quan

In this work, we propose a new framework for the knowledge fusion of chat LLMs through two main stages, resulting in FuseChat.

Instruction Following

Cost-Effective Proxy Reward Model Construction with On-Policy and Active Learning

no code implementations2 Jul 2024 Yifang Chen, Shuohang Wang, ZiYi Yang, Hiteshi Sharma, Nikos Karampatziakis, Donghan Yu, Kevin Jamieson, Simon Shaolei Du, Yelong Shen

Reinforcement learning with human feedback (RLHF), as a widely adopted approach in current large language model pipelines, is \textit{bottlenecked by the size of human preference data}.

Active Learning Language Modelling +2

Self-Exploring Language Models: Active Preference Elicitation for Online Alignment

1 code implementation29 May 2024 Shenao Zhang, Donghan Yu, Hiteshi Sharma, Han Zhong, Zhihan Liu, ZiYi Yang, Shuohang Wang, Hany Hassan, Zhaoran Wang

Preference optimization, particularly through Reinforcement Learning from Human Feedback (RLHF), has achieved significant success in aligning Large Language Models (LLMs) to adhere to human intentions.

Instruction Following

Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone

no code implementations22 Apr 2024 Marah Abdin, Jyoti Aneja, Hany Awadalla, Ahmed Awadallah, Ammar Ahmad Awan, Nguyen Bach, Amit Bahree, Arash Bakhtiari, Jianmin Bao, Harkirat Behl, Alon Benhaim, Misha Bilenko, Johan Bjorck, Sébastien Bubeck, Martin Cai, Qin Cai, Vishrav Chaudhary, Dong Chen, Dongdong Chen, Weizhu Chen, Yen-Chun Chen, Yi-Ling Chen, Hao Cheng, Parul Chopra, Xiyang Dai, Matthew Dixon, Ronen Eldan, Victor Fragoso, Jianfeng Gao, Mei Gao, Min Gao, Amit Garg, Allie Del Giorno, Abhishek Goswami, Suriya Gunasekar, Emman Haider, Junheng Hao, Russell J. Hewett, Wenxiang Hu, Jamie Huynh, Dan Iter, Sam Ade Jacobs, Mojan Javaheripi, Xin Jin, Nikos Karampatziakis, Piero Kauffmann, Mahoud Khademi, Dongwoo Kim, Young Jin Kim, Lev Kurilenko, James R. Lee, Yin Tat Lee, Yuanzhi Li, Yunsheng Li, Chen Liang, Lars Liden, Xihui Lin, Zeqi Lin, Ce Liu, Liyuan Liu, Mengchen Liu, Weishung Liu, Xiaodong Liu, Chong Luo, Piyush Madan, Ali Mahmoudzadeh, David Majercak, Matt Mazzola, Caio César Teodoro Mendes, Arindam Mitra, Hardik Modi, Anh Nguyen, Brandon Norick, Barun Patra, Daniel Perez-Becker, Thomas Portet, Reid Pryzant, Heyang Qin, Marko Radmilac, Liliang Ren, Gustavo de Rosa, Corby Rosset, Sambudha Roy, Olatunji Ruwase, Olli Saarikivi, Amin Saied, Adil Salim, Michael Santacroce, Shital Shah, Ning Shang, Hiteshi Sharma, Yelong Shen, Swadheen Shukla, Xia Song, Masahiro Tanaka, Andrea Tupini, Praneetha Vaddamanu, Chunyu Wang, Guanhua Wang, Lijuan Wang, Shuohang Wang, Xin Wang, Yu Wang, Rachel Ward, Wen Wen, Philipp Witte, Haiping Wu, Xiaoxia Wu, Michael Wyatt, Bin Xiao, Can Xu, Jiahang Xu, Weijian Xu, Jilong Xue, Sonali Yadav, Fan Yang, Jianwei Yang, Yifan Yang, ZiYi Yang, Donghan Yu, Lu Yuan, Chenruidong Zhang, Cyril Zhang, Jianwen Zhang, Li Lyna Zhang, Yi Zhang, Yue Zhang, Yunan Zhang, Xiren Zhou

We introduce phi-3-mini, a 3. 8 billion parameter language model trained on 3. 3 trillion tokens, whose overall performance, as measured by both academic benchmarks and internal testing, rivals that of models such as Mixtral 8x7B and GPT-3. 5 (e. g., phi-3-mini achieves 69% on MMLU and 8. 38 on MT-bench), despite being small enough to be deployed on a phone.

Ranked #5 on MMR total on MRR-Benchmark (using extra training data)

Language Modeling Language Modelling +3

3DGSR: Implicit Surface Reconstruction with 3D Gaussian Splatting

no code implementations30 Mar 2024 Xiaoyang Lyu, Yang-tian Sun, Yi-Hua Huang, Xiuzhe Wu, ZiYi Yang, Yilun Chen, Jiangmiao Pang, Xiaojuan Qi

In this paper, we present an implicit surface reconstruction method with 3D Gaussian Splatting (3DGS), namely 3DGSR, that allows for accurate 3D reconstruction with intricate details while inheriting the high efficiency and rendering quality of 3DGS.

3DGS 3D Reconstruction +1

Knowledge Fusion of Chat LLMs: A Preliminary Technical Report

2 code implementations25 Feb 2024 Fanqi Wan, ZiYi Yang, Longguang Zhong, Xiaojun Quan, Xinting Huang, Wei Bi

Recently, FuseLLM introduced the concept of knowledge fusion to transfer the collective knowledge of multiple structurally varied LLMs into a target LLM through lightweight continual training.

Spec-Gaussian: Anisotropic View-Dependent Appearance for 3D Gaussian Splatting

no code implementations24 Feb 2024 ZiYi Yang, Xinyu Gao, Yangtian Sun, Yihua Huang, Xiaoyang Lyu, Wen Zhou, Shaohui Jiao, Xiaojuan Qi, Xiaogang Jin

Thanks to ASG, we have significantly improved the ability of 3D-GS to model scenes with specular and anisotropic components without increasing the number of 3D Gaussians.

SC-GS: Sparse-Controlled Gaussian Splatting for Editable Dynamic Scenes

1 code implementation CVPR 2024 Yi-Hua Huang, Yang-tian Sun, ZiYi Yang, Xiaoyang Lyu, Yan-Pei Cao, Xiaojuan Qi

During learning, the location and number of control points are adaptively adjusted to accommodate varying motion complexities in different regions, and an ARAP loss following the principle of as rigid as possible is developed to enforce spatial continuity and local rigidity of learned motions.

Novel View Synthesis

CoDi-2: In-Context, Interleaved, and Interactive Any-to-Any Generation

no code implementations30 Nov 2023 Zineng Tang, ZiYi Yang, Mahmoud Khademi, Yang Liu, Chenguang Zhu, Mohit Bansal

We present CoDi-2, a versatile and interactive Multimodal Large Language Model (MLLM) that can follow complex multimodal interleaved instructions, conduct in-context learning (ICL), reason, chat, edit, etc., in an any-to-any input-output modality paradigm.

Image Generation In-Context Learning +5

Soft Convex Quantization: Revisiting Vector Quantization with Convex Optimization

no code implementations4 Oct 2023 Tanmay Gautam, Reid Pryzant, ZiYi Yang, Chenguang Zhu, Somayeh Sojoudi

SCQ works like a differentiable convex optimization (DCO) layer: in the forward pass, we solve for the optimal convex combination of codebook vectors that quantize the inputs.

Image Reconstruction Quantization

Plug in the Safety Chip: Enforcing Constraints for LLM-driven Robot Agents

no code implementations18 Sep 2023 ZiYi Yang, Shreyas S. Raman, Ankit Shah, Stefanie Tellex

Recent advancements in large language models (LLMs) have enabled a new research domain, LLM agents, for solving robotics and planning tasks by leveraging the world knowledge and general reasoning abilities of LLMs obtained during pretraining.

World Knowledge

A General Implicit Framework for Fast NeRF Composition and Rendering

no code implementations9 Aug 2023 Xinyu Gao, ZiYi Yang, Yunlu Zhao, Yuxiang Sun, Xiaogang Jin, Changqing Zou

Mainly, our work introduces a new surface representation known as Neural Depth Fields (NeDF) that quickly determines the spatial relationship between objects by allowing direct intersection computation between rays and implicit surfaces.

NeRF

Multi-task Bioassay Pre-training for Protein-ligand Binding Affinity Prediction

1 code implementation8 Jun 2023 Jiaxian Yan, Zhaofeng Ye, ZiYi Yang, Chengqiang Lu, Shengyu Zhang, Qi Liu, Jiezhong Qiu

By introducing multi-task pre-training to treat the prediction of different affinity labels as different tasks and classifying relative rankings between samples from the same bioassay, MBP learns robust and transferrable structural knowledge from our new ChEMBL-Dock dataset with varied and noisy labels.

Drug Discovery Prediction

i-Code Studio: A Configurable and Composable Framework for Integrative AI

no code implementations23 May 2023 Yuwei Fang, Mahmoud Khademi, Chenguang Zhu, ZiYi Yang, Reid Pryzant, Yichong Xu, Yao Qian, Takuya Yoshioka, Lu Yuan, Michael Zeng, Xuedong Huang

Artificial General Intelligence (AGI) requires comprehensive understanding and generation capabilities for a variety of tasks spanning different modalities and functionalities.

Question Answering Speech-to-Speech Translation +3

i-Code V2: An Autoregressive Generation Framework over Vision, Language, and Speech Data

no code implementations21 May 2023 ZiYi Yang, Mahmoud Khademi, Yichong Xu, Reid Pryzant, Yuwei Fang, Chenguang Zhu, Dongdong Chen, Yao Qian, Mei Gao, Yi-Ling Chen, Robert Gmyr, Naoyuki Kanda, Noel Codella, Bin Xiao, Yu Shi, Lu Yuan, Takuya Yoshioka, Michael Zeng, Xuedong Huang

The convergence of text, visual, and audio data is a key step towards human-like artificial intelligence, however the current Vision-Language-Speech landscape is dominated by encoder-only models which lack generative abilities.

Decoder Diversity

Any-to-Any Generation via Composable Diffusion

2 code implementations NeurIPS 2023 Zineng Tang, ZiYi Yang, Chenguang Zhu, Michael Zeng, Mohit Bansal

We present Composable Diffusion (CoDi), a novel generative model capable of generating any combination of output modalities, such as language, image, video, or audio, from any combination of input modalities.

Ranked #8 on Audio Generation on AudioCaps (FAD metric)

Audio Generation

Real-Time Audio-Visual End-to-End Speech Enhancement

no code implementations13 Mar 2023 Zirun Zhu, Hemin Yang, Min Tang, ZiYi Yang, Sefik Emre Eskimez, Huaming Wang

In this paper, we propose a low-latency real-time audio-visual end-to-end enhancement (AV-E3Net) model based on the recently proposed end-to-end enhancement network (E3Net).

Speech Enhancement Task 2

Grounding Complex Natural Language Commands for Temporal Tasks in Unseen Environments

no code implementations22 Feb 2023 Jason Xinyu Liu, ZiYi Yang, Ifrah Idrees, Sam Liang, Benjamin Schornstein, Stefanie Tellex, Ankit Shah

We propose Lang2LTL, a modular system and a software package that leverages large language models (LLMs) to ground temporal navigational commands to LTL specifications in environments without prior language data.

APOLLO: A Simple Approach for Adaptive Pretraining of Language Models for Logical Reasoning

no code implementations19 Dec 2022 Soumya Sanyal, Yichong Xu, Shuohang Wang, ZiYi Yang, Reid Pryzant, Wenhao Yu, Chenguang Zhu, Xiang Ren

Logical reasoning of text is an important ability that requires understanding the information present in the text, their interconnections, and then reasoning through them to infer new conclusions.

Data Augmentation Language Modeling +4

Unifying Vision, Text, and Layout for Universal Document Processing

5 code implementations CVPR 2023 Zineng Tang, ZiYi Yang, Guoxin Wang, Yuwei Fang, Yang Liu, Chenguang Zhu, Michael Zeng, Cha Zhang, Mohit Bansal

UDOP leverages the spatial correlation between textual content and document image to model image, text, and layout modalities with one uniform representation.

Ranked #5 on Visual Question Answering (VQA) on InfographicVQA (using extra training data)

document understanding Image Reconstruction +1

UniSumm and SummZoo: Unified Model and Diverse Benchmark for Few-Shot Summarization

1 code implementation17 Nov 2022 Yulong Chen, Yang Liu, Ruochen Xu, ZiYi Yang, Chenguang Zhu, Michael Zeng, Yue Zhang

The high annotation costs and diverse demands of various summarization tasks motivate the development of few-shot summarization.

Diversity

MACSum: Controllable Summarization with Mixed Attributes

1 code implementation9 Nov 2022 Yusen Zhang, Yang Liu, ZiYi Yang, Yuwei Fang, Yulong Chen, Dragomir Radev, Chenguang Zhu, Michael Zeng, Rui Zhang

We propose two simple and effective parameter-efficient approaches for the new task of mixed controllable summarization based on hard prompt tuning and soft prefix tuning.

Articles Attribute +1

Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners

1 code implementation22 May 2022 Zhenhailong Wang, Manling Li, Ruochen Xu, Luowei Zhou, Jie Lei, Xudong Lin, Shuohang Wang, ZiYi Yang, Chenguang Zhu, Derek Hoiem, Shih-Fu Chang, Mohit Bansal, Heng Ji

The goal of this work is to build flexible video-language models that can generalize to various video-to-text tasks from few examples, such as domain-specific captioning, question answering, and future event prediction.

Attribute Automatic Speech Recognition +6

ODBO: Bayesian Optimization with Search Space Prescreening for Directed Protein Evolution

1 code implementation19 May 2022 Lixue Cheng, ZiYi Yang, ChangYu Hsieh, Benben Liao, Shengyu Zhang

Directed evolution is a versatile technique in protein engineering that mimics the process of natural selection by iteratively alternating between mutagenesis and screening in order to search for sequences that optimize a given property of interest, such as catalytic activity and binding affinity to a specified target.

Bayesian Optimization Experimental Design +1

Automatic Rule Induction for Interpretable Semi-Supervised Learning

1 code implementation18 May 2022 Reid Pryzant, ZiYi Yang, Yichong Xu, Chenguang Zhu, Michael Zeng

Semi-supervised learning has shown promise in allowing NLP models to generalize from small amounts of labeled data.

Relation Extraction

GASCN: Graph Attention Shape Completion Network

no code implementations20 Jan 2022 Haojie Huang, ZiYi Yang, Robert Platt

Shape completion, the problem of inferring the complete geometry of an object given a partial point cloud, is an important problem in robotics and computer vision.

Graph Attention

SPLDExtraTrees: Robust machine learning approach for predicting kinase inhibitor resistance

no code implementations15 Nov 2021 ZiYi Yang, Zhaofeng Ye, Yijia Xiao, ChangYu Hsieh, Shengyu Zhang

Drug resistance is a major threat to the global health and a significant concern throughout the clinical treatment of diseases and drug development.

BIG-bench Machine Learning

Natural Language for Human-Robot Collaboration: Problems Beyond Language Grounding

no code implementations9 Oct 2021 Seth Pate, Wei Xu, ZiYi Yang, Maxwell Love, Siddarth Ganguri, Lawson L. S. Wong

To enable robots to instruct humans in collaborations, we identify several aspects of language processing that are not commonly studied in this context.

A Simple and Effective Method To Eliminate the Self Language Bias in Multilingual Representations

1 code implementation EMNLP 2021 ZiYi Yang, Yinfei Yang, Daniel Cer, Eric Darve

A simple but highly effective method "Language Information Removal (LIR)" factors out language identity information from semantic related components in multilingual representations pre-trained on multi-monolingual data.

Cross-Lingual Transfer Retrieval

Universal Sentence Representations Learning with Conditional Masked Language Model

no code implementations1 Jan 2021 ZiYi Yang, Yinfei Yang, Daniel M Cer, Jax Law, Eric Darve

This paper presents a novel training method, Conditional Masked Language Modeling (CMLM), to effectively learn sentence representations on large scale unlabeled corpora.

Language Modeling Language Modelling +5

Universal Sentence Representation Learning with Conditional Masked Language Model

no code implementations EMNLP 2021 ZiYi Yang, Yinfei Yang, Daniel Cer, Jax Law, Eric Darve

This paper presents a novel training method, Conditional Masked Language Modeling (CMLM), to effectively learn sentence representations on large scale unlabeled corpora.

Language Modeling Language Modelling +5

Multi-Constitutive Neural Network for Large Deformation Poromechanics Problem

no code implementations11 Oct 2020 Qi Zhang, Yilin Chen, ZiYi Yang, Eric Darve

We propose a novel method "multi-constitutive neural network" (MCNN) such that one model can solve several different constitutive laws.

Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype Prediction

no code implementations2 Sep 2020 Ziyi Yang, Jun Shu, Yong Liang, Deyu Meng, Zongben Xu

Current machine learning has made great progress on computer vision and many other fields attributed to the large amount of high-quality training samples, while it does not work very well on genomic data analysis, since they are notoriously known as small data.

feature selection Few-Shot Image Classification +2

Filtered Inner Product Projection for Crosslingual Embedding Alignment

no code implementations ICLR 2021 Vin Sachidananda, ZiYi Yang, Chenguang Zhu

Due to widespread interest in machine translation and transfer learning, there are numerous algorithms for mapping multiple embeddings to a shared representation space.

Machine Translation Transfer Learning +1

Anomaly Detection with Domain Adaptation

no code implementations5 Jun 2020 Ziyi Yang, Iman Soltani Bozchalooi, Eric Darve

We study the problem of semi-supervised anomaly detection with domain adaptation.

Domain Adaptation Object Recognition +2

Memory Augmented Generative Adversarial Networks for Anomaly Detection

no code implementations7 Feb 2020 Ziyi Yang, Teng Zhang, Iman Soltani Bozchalooi, Eric Darve

Decoded memory units in MEMGAN are more interpretable and disentangled than previous methods, which further demonstrates the effectiveness of the memory mechanism.

Anomaly Detection

Leveraging Lead Bias for Zero-shot Abstractive News Summarization

no code implementations25 Dec 2019 Chenguang Zhu, Ziyi Yang, Robert Gmyr, Michael Zeng, Xuedong Huang

A typical journalistic convention in news articles is to deliver the most salient information in the beginning, also known as the lead bias.

Articles Domain Adaptation +1

Make Lead Bias in Your Favor: A Simple and Effective Method for News Summarization

no code implementations25 Sep 2019 Chenguang Zhu, ZiYi Yang, Robert Gmyr, Michael Zeng, Xuedong Huang

For example, the pretrained model without finetuning outperforms pointer-generator network on CNN/DailyMail dataset.

News Summarization

Embedding Imputation with Grounded Language Information

1 code implementation ACL 2019 Ziyi Yang, Chenguang Zhu, Sachidan, Vin a, Eric Darve

In this paper, we propose an approach for embedding imputation which uses grounded information in the form of a knowledge graph.

Imputation

Out-of-Vocabulary Embedding Imputation with Grounded Language Information by Graph Convolutional Networks

no code implementations ACL 2019 Ziyi Yang, Chenguang Zhu, Vin Sachidananda, Eric Darve

In this paper, we propose an approach for embedding imputation which uses grounded information in the form of a knowledge graph.

Imputation

Parameter-free Sentence Embedding via Orthogonal Basis

1 code implementation IJCNLP 2019 Ziyi Yang, Chenguang Zhu, Weizhu Chen

Inspired by the Gram-Schmidt Process in geometric theory, we build an orthogonal basis of the subspace spanned by a word and its surrounding context in a sentence.

Sentence Sentence Embedding +2

Language Distribution Prediction based on Batch Markov Monte Carlo Simulation with Migration

no code implementations26 Feb 2018 XingYu Fu, ZiYi Yang, XiuWen Duan

To model the randomness of language spreading, we propose the Batch Markov Monte Carlo Simulation with Migration(BMMCSM) algorithm, in which each agent is treated as a language stack.

Cultural Vocal Bursts Intensity Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.