no code implementations • 9 Oct 2024 • Zhengyu Hu, Yichuan Li, Zhengyu Chen, Jingang Wang, Han Liu, Kyumin Lee, Kaize Ding
Textual Attributed Graphs (TAGs) are crucial for modeling complex real-world systems, yet leveraging large language models (LLMs) for TAGs presents unique challenges due to the gap between sequential text processing and graph-structured data.
1 code implementation • 27 Sep 2024 • Wen Ge, Guanyi Mou, Emmanuel O. Agu, Kyumin Lee
Human Activity Recognition (HAR) is a challenging, multi-label classification problem as activities may co-occur and sensor signals corresponding to the same activity may vary in different contexts (e. g., different device placements).
no code implementations • 26 Sep 2024 • Guanyi Mou, Yichuan Li, Kyumin Lee
To boost deep learning models' performance given augmented data/samples in text classification tasks, we propose a novel framework, which leverages both meta learning and contrastive learning techniques as parts of our design for reweighting the augmented samples and refining their feature representations based on their quality.
no code implementations • 26 Sep 2024 • Wen Ge, Guanyi Mou, Emmanuel O. Agu, Kyumin Lee
In this paper, we argue that context-aware activity visit patterns in realistic in-the-wild data can equivocally be considered as a general graph representation learning task.
no code implementations • 25 Sep 2024 • Guanyi Mou, Pengyi Ye, Kyumin Lee
Our model robustly and significantly performed well under extreme adversarial attack (manipulation of 50% messages), achieving 0. 967 accuracy and 0. 934 macro F1.
no code implementations • 25 Sep 2024 • Guanyi Mou, Kyumin Lee
With the widespread online social networks, hate speeches are spreading faster and causing more damage than ever before.
1 code implementation • 25 Sep 2024 • Guanyi Mou, Yun Yue, Kyumin Lee, Ziming Zhang
To counter these environmentally damaging illegal operations, in this research, we focus on wildlife product sales promotion behaviors in online social networks.
no code implementations • 26 Apr 2024 • Yichuan Li, Kaize Ding, Jianling Wang, Kyumin Lee
With the capabilities of understanding and executing natural language instructions, Large language models (LLMs) can potentially act as a powerful tool for textual data augmentation.
1 code implementation • 11 Mar 2024 • Yichuan Li, Xiyao Ma, Sixing Lu, Kyumin Lee, Xiaohu Liu, Chenlei Guo
Large Language models (LLMs) have demonstrated impressive in-context learning (ICL) capabilities, where a LLM makes predictions for a given test input together with a few input-output pairs (demonstrations).
1 code implementation • 23 Oct 2023 • Yichuan Li, Kaize Ding, Kyumin Lee
Self-supervised representation learning on text-attributed graphs, which aims to create expressive and generalizable representations for various downstream tasks, has received increasing research attention lately.
no code implementations • 2 May 2023 • Yichuan Li, Jialong Han, Kyumin Lee, Chengyuan Ma, Benjamin Yao, Derek Liu
In recent years, Pre-trained Language Models (PLMs) have shown their superiority by pre-training on unstructured text corpus and then fine-tuning on downstream tasks.
no code implementations • 17 Jul 2022 • Devin Coughlin, Maylee Gagnon, Victoria Grasso, Guanyi Mou, Kyumin Lee, Renata Konrad, Patricia Raxter, Meredith Gore
Experts combating wildlife trafficking manually sift through articles about seizures and arrests, which is time consuming and make identifying trends difficult.
1 code implementation • EACL 2021 • Nguyen Vo, Kyumin Lee
Our source code and datasets are released at \texttt{\url{https://github. com/nguyenvo09/EACL2021}}.
no code implementations • EMNLP 2020 • Thanh Tran, Yifan Hu, Changwei Hu, Kevin Yen, Fei Tan, Kyumin Lee, Serim Park
HABERTOR inherits BERT's architecture, but is different in four aspects: (i) it generates its own vocabularies and is pre-trained from the scratch using the largest scale hatespeech dataset; (ii) it consists of Quaternion-based factorized components, resulting in a much smaller number of parameters, faster training and inferencing, as well as less memory usage; (iii) it uses our proposed multi-source ensemble heads with a pooling layer for separate input sources, to further enhance its effectiveness; and (iv) it uses a regularized adversarial training with our proposed fine-grained and adaptive noise magnitude to enhance its robustness.
1 code implementation • EMNLP 2020 • Shyam Subramanian, Kyumin Lee
Automated fact extraction and verification is a challenging task that involves finding relevant evidence sentences from a reliable corpus to verify the truthfulness of a claim.
2 code implementations • EMNLP 2020 • Nguyen Vo, Kyumin Lee
The search can directly warn fake news posters and online users (e. g. the posters' followers) about misinformation, discourage them from spreading fake news, and scale up verified content on social media.
no code implementations • 31 Aug 2020 • Thanh Tran, Di You, Kyumin Lee
Quaternion space has brought several benefits over the traditional Euclidean space: Quaternions (i) consist of a real and three imaginary components, encouraging richer representations; (ii) utilize Hamilton product which better encodes the inter-latent interactions across multiple Quaternion components; and (iii) result in a model with smaller degrees of freedom and less prone to overfitting.
1 code implementation • 7 Jan 2020 • Di You, Nguyen Vo, Kyumin Lee, Qiang Liu
To combat fake news, researchers mostly focused on detecting fake news and journalists built and maintained fact-checking sites (e. g., Snopes. com and Politifact. com).
1 code implementation • 5 Oct 2019 • Nguyen Vo, Kyumin Lee
In fighting against fake news, many fact-checking systems comprised of human-based fact-checking sites (e. g., snopes. com and politifact. com) and automatic detection systems have been developed in recent years.
1 code implementation • 8 Jun 2019 • Thanh Tran, Renee Sweeney, Kyumin Lee
Our first approach uses user-playlist-song interactions, and combines Mahalanobis distance scores between (i) a target user and a target song, and (ii) between a target playlist and the target song to account for both the user's preference and the playlist's theme.
1 code implementation • 1 May 2019 • Thanh Tran, Xinyue Liu, Kyumin Lee, Xiangnan Kong
Personalized recommendation algorithms learn a user's preference for an item by measuring a distance/similarity between them.
2 code implementations • 31 Aug 2018 • Thanh Tran, Kyumin Lee, Yiming Liao, Dongwon Lee
Following recent successes in exploiting both latent factor and word embedding models in recommendation, we propose a novel Regularized Multi-Embedding (RME) based recommendation model that simultaneously encapsulates the following ideas via decomposition: (1) which items a user likes, (2) which two users co-like the same items, (3) which two items users often co-liked, and (4) which two items users often co-disliked.
1 code implementation • 20 Jun 2018 • Nguyen Vo, Kyumin Lee
Despite the existence of these systems, fake news is still wildly shared by online users.
Information Retrieval Social and Information Networks
no code implementations • 12 Oct 2017 • Thanh Tran, Kyumin Lee, Nguyen Vo, Hongkyu Choi
In Crowdfunding platforms, people turn their prototype ideas into real products by raising money from the crowd, or invest in someone else's projects.