Search Results for author: Zixia Jia

Found 15 papers, 9 papers with code

SHARP: Search-Based Adversarial Attack for Structured Prediction

no code implementations Findings (NAACL) 2022 Liwen Zhang, Zixia Jia, Wenjuan Han, Zilong Zheng, Kewei Tu

Adversarial attack of structured prediction models faces various challenges such as the difficulty of perturbing discrete words, the sentence quality issue, and the sensitivity of outputs to small perturbations.

Adversarial Attack Dependency Parsing +5

Seek in the Dark: Reasoning via Test-Time Instance-Level Policy Gradient in Latent Space

no code implementations19 May 2025 Hengli Li, Chenxi Li, Tong Wu, Xuekai Zhu, Yuxuan Wang, Zhaoxin Yu, Eric Hanchen Jiang, Song-Chun Zhu, Zixia Jia, Ying Nian Wu, Zilong Zheng

We introduce LatentSeek, a novel framework that enhances LLM reasoning through Test-Time Instance-level Adaptation (TTIA) within the model's latent space.

GSM8K Math

From Hours to Minutes: Lossless Acceleration of Ultra Long Sequence Generation up to 100K Tokens

1 code implementation26 Feb 2025 Tong Wu, Junzhe Shen, Zixia Jia, Yuxuan Wang, Zilong Zheng

While traditional speculative decoding methods exist, simply extending their generation limits fails to accelerate the process and can be detrimental.

Sparser is Faster and Less is More: Efficient Sparse Attention for Long-Range Transformers

no code implementations24 Jun 2024 Chao Lou, Zixia Jia, Zilong Zheng, Kewei Tu

Accommodating long sequences efficiently in autoregressive Transformers, especially within an extended context window, poses significant challenges due to the quadratic computational complexity and substantial KV memory requirements inherent in self-attention mechanisms.

Language Modeling Language Modelling

LangSuitE: Planning, Controlling and Interacting with Large Language Models in Embodied Text Environments

1 code implementation24 Jun 2024 Zixia Jia, Mengmeng Wang, Baichen Tong, Song-Chun Zhu, Zilong Zheng

Recent advances in Large Language Models (LLMs) have shown inspiring achievements in constructing autonomous agents that rely on language descriptions as inputs.

World Knowledge

Semi-automatic Data Enhancement for Document-Level Relation Extraction with Distant Supervision from Large Language Models

1 code implementation13 Nov 2023 Junpeng Li, Zixia Jia, Zilong Zheng

Document-level Relation Extraction (DocRE), which aims to extract relations from a long context, is a critical challenge in achieving fine-grained structural comprehension and generating interpretable document representations.

Document-level Relation Extraction In-Context Learning +5

Modeling Instance Interactions for Joint Information Extraction with Neural High-Order Conditional Random Field

1 code implementation17 Dec 2022 Zixia Jia, Zhaohui Yan, Wenjuan Han, Zilong Zheng, Kewei Tu

Prior works on joint Information Extraction (IE) typically model instance (e. g., event triggers, entities, roles, relations) interactions by representation enhancement, type dependencies scoring, or global decoding.

Decoder Variational Inference

ITA: Image-Text Alignments for Multi-Modal Named Entity Recognition

1 code implementation NAACL 2022 Xinyu Wang, Min Gui, Yong Jiang, Zixia Jia, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu

As text representations take the most important role in MNER, in this paper, we propose {\bf I}mage-{\bf t}ext {\bf A}lignments (ITA) to align image features into the textual space, so that the attention mechanism in transformer-based pretrained textual embeddings can be better utilized.

Multi-modal Named Entity Recognition named-entity-recognition +1

Semi-Supervised Semantic Dependency Parsing Using CRF Autoencoders

1 code implementation ACL 2020 Zixia Jia, Youmi Ma, Jiong Cai, Kewei Tu

Semantic dependency parsing, which aims to find rich bi-lexical relationships, allows words to have multiple dependency heads, resulting in graph-structured representations.

ARC Decoder +3

Cannot find the paper you are looking for? You can Submit a new open access paper.