no code implementations • 4 Mar 2024 • Fiona Anting Tan, Gerard Christopher Yeo, Fanyou Wu, Weijie Xu, Vinija Jain, Aman Chadha, Kokil Jaidka, Yang Liu, See-Kiong Ng
Drawing inspiration from psychological research on the links between certain personality traits and Theory-of-Mind (ToM) reasoning, and from prompt engineering research on the hyper-sensitivity of prompts in affecting LLMs capabilities, this study investigates how inducing personalities in LLMs using prompts affects their ToM reasoning capabilities.
no code implementations • 28 Oct 2023 • Yixin Wan, Fanyou Wu, Weijie Xu, Srinivasan H. Sengamedu
We explore the correlation between the level of hallucination in model responses and two types of sequence-level certainty: probabilistic certainty and semantic certainty.
1 code implementation • 23 Oct 2023 • Weijie Xu, Wenxiang Hu, Fanyou Wu, Srinivasan Sengamedu
Additionally, by exploiting the power of diffusion model, our framework also provides the capability to do topic based text generation.
1 code implementation • 30 Jul 2022 • Fanyou Wu, Yang Liu, Rado Gazo, Benes Bedrich, Xiaobo Qu
In the Amazon KDD Cup 2022, we aim to apply natural language processing methods to improve the quality of search results that can significantly enhance user experience and engagement with search engines for e-commerce.
no code implementations • 13 Nov 2020 • Fanyou Wu, Yang Liu, Zhiyuan Liu, Xiaobo Qu, Rado Gazo, Eva Haviarova
In our 2020 Competition solution, we further design multiple variants based on HR-NET and UNet.
1 code implementation • 11 Nov 2019 • Yang Liu, Fanyou Wu, Baosheng Yu, Zhiyuan Liu, Jieping Ye
How to build an effective large-scale traffic state prediction system is a challenging but highly valuable problem.
1 code implementation • 7 Jun 2019 • Fanyou Wu, Rado Gazo, Eva Haviarova, Bedrich Benes
Consider $l_2$ norms attacks, Project Gradient Descent (PGD) and the Carlini and Wagner (C\&W) attacks are the two main methods, where PGD control max perturbation for adversarial examples while C\&W approach treats perturbation as a regularization term optimized it with loss function together.