Search Results for author: Sungchul Kim

Found 65 papers, 14 papers with code

Few-Shot Class-Incremental Learning for Named Entity Recognition

no code implementations ACL 2022 Rui Wang, Tong Yu, Handong Zhao, Sungchul Kim, Subrata Mitra, Ruiyi Zhang, Ricardo Henao

In this work, we study a more challenging but practical problem, i. e., few-shot class-incremental learning for NER, where an NER model is trained with only few labeled samples of the new classes, without forgetting knowledge of the old ones.

class-incremental learning Few-Shot Class-Incremental Learning +4

Improving the performance of optical inverse design of multilayer thin films using CNN-LSTM tandem neural networks

no code implementations11 Jun 2025 Uijun Jung, Deokho Jang, Sungchul Kim, Jungho Kim

In this paper, we propose to use not only MLP, but also convolutional neural network (CNN) or long short-term memory (LSTM) algorithms in the configuration of the TNN.

WaterFlow: Learning Fast & Robust Watermarks using Stable Diffusion

no code implementations15 Apr 2025 Vinay Shukla, Prachee Sharma, Ryan Rossi, Sungchul Kim, Tong Yu, Aditya Grover

In this work, we propose WaterFlow (WF), a fast and extremely robust approach for high fidelity visual watermarking based on a learned latent-dependent watermark.

Training Robust Graph Neural Networks by Modeling Noise Dependencies

1 code implementation27 Feb 2025 Yeonjun In, Kanghoon Yoon, Sukwon Yun, Kibum Kim, Sungchul Kim, Chanyoung Park

In real-world applications, node features in graphs often contain noise from various sources, leading to significant performance degradation in GNNs.

Variational Inference

From Documents to Dialogue: Building KG-RAG Enhanced AI Assistants

no code implementations21 Feb 2025 Manisha Mukherjee, Sungchul Kim, Xiang Chen, Dan Luo, Tong Yu, Tung Mai

The Adobe Experience Platform AI Assistant is a conversational tool that enables organizations to interact seamlessly with proprietary enterprise data through a chatbot.

Chatbot Entity Resolution +2

Mitigating Visual Knowledge Forgetting in MLLM Instruction-tuning via Modality-decoupled Gradient Descent

no code implementations17 Feb 2025 Junda Wu, Yuxin Xiong, Xintong Li, Yu Xia, Ruoyu Wang, Yu Wang, Tong Yu, Sungchul Kim, Ryan A. Rossi, Lina Yao, Jingbo Shang, Julian McAuley

By explicitly disentangling the optimization of visual understanding from task-specific alignment, MDGD preserves pre-trained visual knowledge while enabling efficient task adaptation.

Continual Learning parameter-efficient fine-tuning

Interactive Visualization Recommendation with Hier-SUCB

no code implementations5 Feb 2025 Songwen Hu, Ryan A. Rossi, Tong Yu, Junda Wu, Handong Zhao, Sungchul Kim, Shuai Li

For more interactive and accurate recommendations, we propose Hier-SUCB, a contextual combinatorial semi-bandit in the PVisRec setting.

Understanding How Paper Writers Use AI-Generated Captions in Figure Caption Writing

no code implementations10 Jan 2025 Ho Yin, Ng, Ting-Yao Hsu, Jiyoo Min, Sungchul Kim, Ryan A. Rossi, Tong Yu, Hyunggu Jung, Ting-Hao 'Kenneth' Huang

By analyzing video recordings of the writing process through interaction analysis, we observed that participants often began by copying and refining AI-generated captions.

Caption Generation

VipAct: Visual-Perception Enhancement via Specialized VLM Agent Collaboration and Tool-use

no code implementations21 Oct 2024 Zhehao Zhang, Ryan Rossi, Tong Yu, Franck Dernoncourt, Ruiyi Zhang, Jiuxiang Gu, Sungchul Kim, Xiang Chen, Zichao Wang, Nedim Lipka

In this paper, we present VipAct, an agent framework that enhances VLMs by integrating multi-agent collaboration and vision expert models, enabling more precise visual understanding and comprehensive reasoning.

Image Captioning Task Planning

Knowledge-Aware Query Expansion with Large Language Models for Textual and Relational Retrieval

no code implementations17 Oct 2024 Yu Xia, Junda Wu, Sungchul Kim, Tong Yu, Ryan A. Rossi, Haoliang Wang, Julian McAuley

Large language models (LLMs) have been used to generate query expansions augmenting original queries for improving information search.

Retrieval

A Multi-LLM Debiasing Framework

no code implementations20 Sep 2024 Deonna M. Owens, Ryan A. Rossi, Sungchul Kim, Tong Yu, Franck Dernoncourt, Xiang Chen, Ruiyi Zhang, Jiuxiang Gu, Hanieh Deilamsalehy, Nedim Lipka

Large Language Models (LLMs) are powerful tools with the potential to benefit society immensely, yet, they have demonstrated biases that perpetuate societal inequalities.

Data Augmentation Human Detection

Visual Prompting in Multimodal Large Language Models: A Survey

no code implementations5 Sep 2024 Junda Wu, Zhehao Zhang, Yu Xia, Xintong Li, Zhaoyang Xia, Aaron Chang, Tong Yu, Sungchul Kim, Ryan A. Rossi, Ruiyi Zhang, Subrata Mitra, Dimitris N. Metaxas, Lina Yao, Jingbo Shang, Julian McAuley

This paper presents the first comprehensive survey on visual prompting methods in MLLMs, focusing on visual prompting, prompt generation, compositional reasoning, and prompt learning.

In-Context Learning Prompt Learning +3

Diversify-verify-adapt: Efficient and Robust Retrieval-Augmented Ambiguous Question Answering

no code implementations4 Sep 2024 Yeonjun In, Sungchul Kim, Ryan A. Rossi, Md Mehrab Tanjim, Tong Yu, Ritwik Sinha, Chanyoung Park

The retrieval augmented generation (RAG) framework addresses an ambiguity in user queries in QA systems by retrieving passages that cover all plausible interpretations and generating comprehensive responses based on the passages.

Question Answering RAG +2

Learning to Reduce: Towards Improving Performance of Large Language Models on Structured Data

no code implementations3 Jul 2024 Younghun Lee, Sungchul Kim, Ryan A. Rossi, Tong Yu, Xiang Chen

This paper proposes a framework, Learning to Reduce, that fine-tunes a language model with On-Policy Learning to generate a reduced version of an input structured data.

Language Modeling Language Modelling

Hallucination Diversity-Aware Active Learning for Text Summarization

no code implementations2 Apr 2024 Yu Xia, Xu Liu, Tong Yu, Sungchul Kim, Ryan A. Rossi, Anup Rao, Tung Mai, Shuai Li

Large Language Models (LLMs) have shown propensity to generate hallucinated outputs, i. e., texts that are factually incorrect or unsupported.

Active Learning Diversity +2

Which LLM to Play? Convergence-Aware Online Model Selection with Time-Increasing Bandits

no code implementations11 Mar 2024 Yu Xia, Fang Kong, Tong Yu, Liya Guo, Ryan A. Rossi, Sungchul Kim, Shuai Li

In this paper, we propose a time-increasing bandit algorithm TI-UCB, which effectively predicts the increase of model performances due to finetuning and efficiently balances exploration and exploitation in model selection.

Change Detection Model Selection

Learning to Reduce: Optimal Representations of Structured Data in Prompting Large Language Models

no code implementations22 Feb 2024 Younghun Lee, Sungchul Kim, Tong Yu, Ryan A. Rossi, Xiang Chen

The model learns to reduce the input context using On-Policy Reinforcement Learning and aims to improve the reasoning performance of a fixed LLM.

Language Modeling Language Modelling

Self-Debiasing Large Language Models: Zero-Shot Recognition and Reduction of Stereotypes

no code implementations3 Feb 2024 Isabel O. Gallegos, Ryan A. Rossi, Joe Barrow, Md Mehrab Tanjim, Tong Yu, Hanieh Deilamsalehy, Ruiyi Zhang, Sungchul Kim, Franck Dernoncourt

Large language models (LLMs) have shown remarkable advances in language generation and understanding but are also prone to exhibiting harmful social biases.

Text Generation Zero-Shot Learning

Augment before You Try: Knowledge-Enhanced Table Question Answering via Table Expansion

1 code implementation28 Jan 2024 Yujian Liu, Jiabao Ji, Tong Yu, Ryan Rossi, Sungchul Kim, Handong Zhao, Ritwik Sinha, Yang Zhang, Shiyu Chang

Table question answering is a popular task that assesses a model's ability to understand and interact with structured data.

Question Answering

GPT-4 as an Effective Zero-Shot Evaluator for Scientific Figure Captions

no code implementations23 Oct 2023 Ting-Yao Hsu, Chieh-Yang Huang, Ryan Rossi, Sungchul Kim, C. Lee Giles, Ting-Hao K. Huang

We first constructed SCICAP-EVAL, a human evaluation dataset that contains human judgments for 3, 600 scientific figure captions, both original and machine-made, for 600 arXiv figures.

Bias and Fairness in Large Language Models: A Survey

1 code implementation2 Sep 2023 Isabel O. Gallegos, Ryan A. Rossi, Joe Barrow, Md Mehrab Tanjim, Sungchul Kim, Franck Dernoncourt, Tong Yu, Ruiyi Zhang, Nesreen K. Ahmed

Rapid advancements of large language models (LLMs) have enabled the processing, understanding, and generation of human-like text, with increasing integration into systems that touch our social sphere.

counterfactual Fairness +1

Structured Dynamic Pricing: Optimal Regret in a Global Shrinkage Model

no code implementations28 Mar 2023 Rashmi Ranjan Bhuyan, Adel Javanmard, Sungchul Kim, Gourab Mukherjee, Ryan A. Rossi, Tong Yu, Handong Zhao

We consider dynamic pricing strategies in a streamed longitudinal data set-up where the objective is to maximize, over time, the cumulative profit across a large number of customer segments.

Graph Learning with Localized Neighborhood Fairness

no code implementations22 Dec 2022 April Chen, Ryan Rossi, Nedim Lipka, Jane Hoffswell, Gromit Chan, Shunan Guo, Eunyee Koh, Sungchul Kim, Nesreen K. Ahmed

Learning fair graph representations for downstream applications is becoming increasingly important, but existing work has mostly focused on improving fairness at the global level by either modifying the graph structure or objective function without taking into account the local neighborhood of a node.

Fairness Graph Learning +2

Direct Embedding of Temporal Network Edges via Time-Decayed Line Graphs

no code implementations30 Sep 2022 Sudhanshu Chanpuriya, Ryan A. Rossi, Sungchul Kim, Tong Yu, Jane Hoffswell, Nedim Lipka, Shunan Guo, Cameron Musco

We present a simple method that avoids both shortcomings: construct the line graph of the network, which includes a node for each interaction, and weigh the edges of this graph based on the difference in time between interactions.

Edge Classification Link Prediction

Bundle MCR: Towards Conversational Bundle Recommendation

1 code implementation26 Jul 2022 Zhankui He, Handong Zhao, Tong Yu, Sungchul Kim, Fan Du, Julian McAuley

MCR, which uses a conversational paradigm to elicit user interests by asking user preferences on tags (e. g., categories or attributes) and handling user feedback across multiple rounds, is an emerging recommendation setting to acquire user feedback and narrow down the output space, but has not been explored in the context of bundle recommendation.

Conversational Recommendation Recommendation Systems

CGC: Contrastive Graph Clustering for Community Detection and Tracking

1 code implementation5 Apr 2022 Namyong Park, Ryan Rossi, Eunyee Koh, Iftikhar Ahamath Burhanuddin, Sungchul Kim, Fan Du, Nesreen Ahmed, Christos Faloutsos

Especially, deep graph clustering (DGC) methods have successfully extended deep clustering to graph-structured data by learning node representations and cluster assignments in a joint optimization framework.

Clustering Community Detection +4

Automatic Forecasting via Meta-Learning

no code implementations29 Sep 2021 Mustafa Abdallah, Ryan Rossi, Kanak Mahadik, Sungchul Kim, Handong Zhao, Haoliang Wang, Saurabh Bagchi

In this work, we develop techniques for fast automatic selection of the best forecasting model for a new unseen time-series dataset, without having to first train (or evaluate) all the models on the new time-series data to select the best one.

Meta-Learning Time Series +1

Influence-guided Data Augmentation for Neural Tensor Completion

1 code implementation23 Aug 2021 Sejoon Oh, Sungchul Kim, Ryan A. Rossi, Srijan Kumar

In this paper, we propose DAIN, a general data augmentation framework that enhances the prediction accuracy of neural tensor completion methods.

Data Augmentation Imputation +3

Personalized Visualization Recommendation

no code implementations12 Feb 2021 Xin Qian, Ryan A. Rossi, Fan Du, Sungchul Kim, Eunyee Koh, Sana Malik, Tak Yeon Lee, Nesreen K. Ahmed

Visualization recommendation work has focused solely on scoring visualizations based on the underlying dataset and not the actual user and their past visualization feedback.

Learning Contextualized Knowledge Graph Structures for Commonsense Reasoning

no code implementations1 Jan 2021 Jun Yan, Mrigank Raman, Tianyu Zhang, Ryan Rossi, Handong Zhao, Sungchul Kim, Nedim Lipka, Xiang Ren

Recently, neural-symbolic architectures have achieved success on commonsense reasoning through effectively encoding relational structures retrieved from external knowledge graphs (KGs) and obtained state-of-the-art results in tasks such as (commonsense) question answering and natural language inference.

Knowledge Graphs Natural Language Inference +1

Heterogeneous Graphlets

no code implementations23 Oct 2020 Ryan A. Rossi, Nesreen K. Ahmed, Aldo Carranza, David Arbour, Anup Rao, Sungchul Kim, Eunyee Koh

Notably, since typed graphlet is more general than colored graphlet (and untyped graphlets), the counts of various typed graphlets can be combined to obtain the counts of the much simpler notion of colored graphlets.

Graph Deep Factors for Forecasting

no code implementations14 Oct 2020 Hongjie Chen, Ryan A. Rossi, Kanak Mahadik, Sungchul Kim, Hoda Eldardiry

GraphDF is a hybrid forecasting framework that consists of a relational global and relational local model.

Computational Efficiency Time Series +1

ML-based Visualization Recommendation: Learning to Recommend Visualizations from Data

no code implementations25 Sep 2020 Xin Qian, Ryan A. Rossi, Fan Du, Sungchul Kim, Eunyee Koh, Sana Malik, Tak Yeon Lee, Joel Chan

Finally, we observed a strong preference by the human experts in our user study towards the visualizations recommended by our ML-based system as opposed to the rule-based system (5. 92 from a 7-point Likert scale compared to only 3. 45).

From Static to Dynamic Node Embeddings

no code implementations21 Sep 2020 Di Jin, Sungchul Kim, Ryan A. Rossi, Danai Koutra

While previous work on dynamic modeling and embedding has focused on representing a stream of timestamped edges using a time-series of graphs based on a specific time-scale (e. g., 1 month), we propose the notion of an $\epsilon$-graph time-series that uses a fixed number of edges for each graph, and show its superiority over the time-scale representation used in previous work.

Time Series Time Series Analysis

On Proximity and Structural Role-based Embeddings in Networks: Misconceptions, Techniques, and Applications

no code implementations22 Aug 2019 Ryan A. Rossi, Di Jin, Sungchul Kim, Nesreen K. Ahmed, Danai Koutra, John Boaz Lee

Unfortunately, recent work has sometimes confused the notion of structural roles and communities (based on proximity) leading to misleading or incorrect claims about the capabilities of network embedding methods.

Misconceptions Network Embedding

Higher-Order Ranking and Link Prediction: From Closing Triangles to Closing Higher-Order Motifs

no code implementations12 Jun 2019 Ryan A. Rossi, Anup Rao, Sungchul Kim, Eunyee Koh, Nesreen K. Ahmed, Gang Wu

In this work, we investigate higher-order network motifs and develop techniques based on the notion of closing higher-order motifs that move beyond closing simple triangles.

Link Prediction Prediction

Figure Captioning with Reasoning and Sequence-Level Training

no code implementations7 Jun 2019 Charles Chen, Ruiyi Zhang, Eunyee Koh, Sungchul Kim, Scott Cohen, Tong Yu, Ryan Rossi, Razvan Bunescu

In this work, we investigate the problem of figure captioning where the goal is to automatically generate a natural language description of the figure.

Image Captioning Reinforcement Learning

Dynamic Node Embeddings from Edge Streams

no code implementations12 Apr 2019 John Boaz Lee, Giang Nguyen, Ryan A. Rossi, Nesreen K. Ahmed, Eunyee Koh, Sungchul Kim

In this work, we propose using the notion of temporal walks for learning dynamic embeddings from temporal networks.

Representation Learning valid

Heterogeneous Network Motifs

no code implementations28 Jan 2019 Ryan A. Rossi, Nesreen K. Ahmed, Aldo Carranza, David Arbour, Anup Rao, Sungchul Kim, Eunyee Koh

To address this problem, we propose a fast, parallel, and space-efficient framework for counting typed graphlets in large networks.

Latent Network Summarization: Bridging Network Embedding and Summarization

1 code implementation11 Nov 2018 Di Jin, Ryan Rossi, Danai Koutra, Eunyee Koh, Sungchul Kim, Anup Rao

Motivated by the computational and storage challenges that dense embeddings pose, we introduce the problem of latent network summarization that aims to learn a compact, latent representation of the graph structure with dimensionality that is independent of the input graph size (i. e., #nodes and #edges), while retaining the ability to derive node representations on the fly.

Social and Information Networks

Higher-order Graph Convolutional Networks

no code implementations12 Sep 2018 John Boaz Lee, Ryan A. Rossi, Xiangnan Kong, Sungchul Kim, Eunyee Koh, Anup Rao

Experiments show that our proposed method is able to achieve state-of-the-art results on the semi-supervised node classification task.

General Classification Graph Attention +1

Attention Models in Graphs: A Survey

1 code implementation20 Jul 2018 John Boaz Lee, Ryan A. Rossi, Sungchul Kim, Nesreen K. Ahmed, Eunyee Koh

However, in the real-world, graphs can be both large - with many complex patterns - and noisy which can pose a problem for effective graph mining.

Graph Attention Graph Classification +3

HONE: Higher-Order Network Embeddings

no code implementations28 Jan 2018 Ryan A. Rossi, Nesreen K. Ahmed, Eunyee Koh, Sungchul Kim, Anup Rao, Yasin Abbasi Yadkori

This paper describes a general framework for learning Higher-Order Network Embeddings (HONE) from graph data based on network motifs.

Cannot find the paper you are looking for? You can Submit a new open access paper.