no code implementations • 15 Nov 2024 • Wenxuan Wang, Wenxiang Jiao, Jen-tse Huang, Zhaopeng Tu, Michael R. Lyu
By carefully designing experiments on different MNMT scenarios and models, we attribute the off-target issue to the overfitting of the shortcuts of (non-centric, centric) language mappings.
1 code implementation • 5 Nov 2024 • Jingyu Xiao, Yuxuan Wan, Yintong Huo, Zhiyao Xu, Michael R. Lyu
To fill in the blank, we present the first systematic investigation of MLLMs in generating interactive webpages.
1 code implementation • 8 Oct 2024 • Zi-Yuan Hu, Yiwu Zhong, Shijia Huang, Michael R. Lyu, LiWei Wang
However, most existing Video LLMs neglect temporal information in video data, leading to struggles with temporal-aware video understanding.
1 code implementation • 20 Sep 2024 • JunJie Huang, Daya Guo, Chenglong Wang, Jiazhen Gu, Shuai Lu, Jeevana Priya Inala, Cong Yan, Jianfeng Gao, Nan Duan, Michael R. Lyu
With CoCoMine, we construct CoCoNote, a dataset containing 58, 221 examples for Contextualized Data-wrangling Code generation in Notebooks.
1 code implementation • 20 Sep 2024 • JunJie Huang, Zhihan Jiang, Jinyang Liu, Yintong Huo, Jiazhen Gu, Zhuangbin Chen, Cong Feng, Hui Dong, Zengyin Yang, Michael R. Lyu
In the second stage, LoFI leverages a pre-trained language model with a novel prompt-based tuning method to extract fine-grained information of interest from the collected logs.
no code implementations • 31 Aug 2024 • Wenxuan Wang, Juluan Shi, Chaozheng Wang, Cheryl Lee, Youliang Yuan, Jen-tse Huang, Michael R. Lyu
Equipped with the capability to call functions, modern large language models (LLMs) can leverage external tools for addressing a range of tasks unattainable through language skills alone.
1 code implementation • 6 Aug 2024 • Yanyang Li, Shuo Liang, Michael R. Lyu, LiWei Wang
Recent advancements in long-context modeling have enhanced language models (LMs) for complex tasks across multiple NLP applications.
1 code implementation • 2 Aug 2024 • Jen-tse Huang, Jiaxu Zhou, Tailin Jin, Xuhui Zhou, Zixi Chen, Wenxuan Wang, Youliang Yuan, Maarten Sap, Michael R. Lyu
Additionally, we show the promise of improving multi-agent system resilience by demonstrating that two defense methods, introducing a mechanism for each agent to challenge others' outputs, or an additional agent to review and correct messages, can enhance system resilience.
no code implementations • 24 Jun 2024 • Yuxuan Wan, Chaozheng Wang, Yi Dong, Wenxuan Wang, Shuqing Li, Yintong Huo, Michael R. Lyu
We conduct extensive testing with a dataset comprised of real-world websites and various MLLMs and demonstrate that DCGen achieves up to a 14% improvement in visual similarity over competing methods.
no code implementations • 3 May 2024 • Michael R. Lyu, Baishakhi Ray, Abhik Roychoudhury, Shin Hwei Tan, Patanamon Thongtanunam
In this article, we study automated coding in a general sense and study the concerns around code quality, security and related issues of programmer responsibility.
1 code implementation • 22 Apr 2024 • Man Tik Ng, Hui Tung Tse, Jen-tse Huang, Jingjing Li, Wenxuan Wang, Michael R. Lyu
However, existing studies focus on imitating well-known public figures or fictional characters, overlooking the potential for simulating ordinary individuals.
1 code implementation • 27 Mar 2024 • Yiwu Zhong, Zi-Yuan Hu, Michael R. Lyu, LiWei Wang
Visual representation learning has been a cornerstone in computer vision, involving typical forms such as visual embeddings, structural symbols, and text-based representations.
Ranked #85 on Visual Question Answering on MM-Vet
1 code implementation • 18 Mar 2024 • Jen-tse Huang, Eric John Li, Man Ho Lam, Tian Liang, Wenxuan Wang, Youliang Yuan, Wenxiang Jiao, Xing Wang, Zhaopeng Tu, Michael R. Lyu
Researchers have examined LLMs' decision-making through the lens of Game Theory.
no code implementations • 11 Mar 2024 • Jinxi Kuang, Jinyang Liu, JunJie Huang, Renyi Zhong, Jiazhen Gu, Lan Yu, Rui Tan, Zengyin Yang, Michael R. Lyu
We also share our experience in deploying COLA in our real-world cloud system, Cloud X.
no code implementations • 27 Feb 2024 • JunJie Huang, Jinyang Liu, Zhuangbin Chen, Zhihan Jiang, Yichen Li, Jiazhen Gu, Cong Feng, Zengyin Yang, Yongqiang Yang, Michael R. Lyu
To date, FaultProfIT has analyzed 10, 000+ incidents from 30+ cloud services, successfully revealing several fault trends that have informed system improvements.
no code implementations • 17 Feb 2024 • Jie Liu, Wenxuan Wang, Yihang Su, Jingyuan Huan, WenTing Chen, Yudi Zhang, Cheng-Yi Li, Kao-Jung Chang, Xiaohan Xin, Linlin Shen, Michael R. Lyu
The significant breakthroughs of Medical Multi-Modal Large Language Models (Med-MLLMs) renovate modern healthcare with robust information synthesis and medical decision support.
no code implementations • 6 Feb 2024 • Yichen Li, Yun Peng, Yintong Huo, Michael R. Lyu
We conducted preliminary experiments to validate the performance of IDECoder and observed that this synergy represents a promising trend for future exploration.
1 code implementation • 10 Jan 2024 • Jinyang Liu, Wenwei Gu, Zhuangbin Chen, Yichen Li, Yuxin Su, Michael R. Lyu
These methods are evaluated with five multivariate KPI datasets that are publicly available.
no code implementations • 1 Jan 2024 • Wenxuan Wang, Haonan Bai, Jen-tse Huang, Yuxuan Wan, Youliang Yuan, Haoyi Qiu, Nanyun Peng, Michael R. Lyu
BiasPainter uses a diverse range of seed images of individuals and prompts the image generation models to edit these images using gender, race, and age-neutral queries.
1 code implementation • 1 Jan 2024 • Yuxuan Wan, Wenxuan Wang, Yiliu Yang, Youliang Yuan, Jen-tse Huang, Pinjia He, Wenxiang Jiao, Michael R. Lyu
We introduce LogicAsker, a novel approach for evaluating and enhancing the logical reasoning capabilities of large language models (LLMs) such as ChatGPT and GPT-4.
no code implementations • 1 Jan 2024 • Wenxuan Wang, Juluan Shi, Zhaopeng Tu, Youliang Yuan, Jen-tse Huang, Wenxiang Jiao, Michael R. Lyu
Current methods for evaluating LLMs' veracity are limited by test data leakage or the need for extensive human labor, hindering efficient and accurate error detection.
no code implementations • 19 Oct 2023 • Wenxuan Wang, Wenxiang Jiao, Jingyuan Huang, Ruyi Dai, Jen-tse Huang, Zhaopeng Tu, Michael R. Lyu
This paper identifies a cultural dominance issue within large language models (LLMs) due to the predominant use of English data in model training (e. g., ChatGPT).
1 code implementation • 2 Oct 2023 • Wenxuan Wang, Zhaopeng Tu, Chang Chen, Youliang Yuan, Jen-tse Huang, Wenxiang Jiao, Michael R. Lyu
In this work, we build the first multilingual safety benchmark for LLMs, XSafety, in response to the global deployment of LLMs in practice.
1 code implementation • 2 Oct 2023 • Jen-tse Huang, Wenxuan Wang, Eric John Li, Man Ho Lam, Shujie Ren, Youliang Yuan, Wenxiang Jiao, Zhaopeng Tu, Michael R. Lyu
Large Language Models (LLMs) have recently showcased their remarkable capacities, not only in natural language processing tasks but also across diverse domains such as clinical medicine, legal consultation, and education.
no code implementations • 19 Aug 2023 • Jinyang Liu, Tianyi Yang, Zhuangbin Chen, Yuxin Su, Cong Feng, Zengyin Yang, Michael R. Lyu
As modern software systems continue to grow in terms of complexity and volume, anomaly detection on multivariate monitoring metrics, which profile systems' health status, becomes more and more critical and challenging.
no code implementations • 18 Aug 2023 • Wenxuan Wang, Jingyuan Huang, Jen-tse Huang, Chang Chen, Jiazhen Gu, Pinjia He, Michael R. Lyu
Moreover, through retraining the models with the test cases generated by OASIS, the robustness of the moderation model can be improved without performance degradation.
1 code implementation • ICCV 2023 • Zi-Yuan Hu, Yanyang Li, Michael R. Lyu, LiWei Wang
In particular, our VL-PET-large with lightweight PET module designs significantly outperforms VL-Adapter by 2. 92% (3. 41%) and LoRA by 3. 37% (7. 03%) with BART-base (T5-base) on image-text tasks.
1 code implementation • 9 Aug 2023 • Yanyang Li, Jianqiao Zhao, Duo Zheng, Zi-Yuan Hu, Zhi Chen, Xiaohui Su, Yongfeng Huang, Shijia Huang, Dahua Lin, Michael R. Lyu, LiWei Wang
With the continuous emergence of Chinese Large Language Models (LLMs), how to evaluate a model's capabilities has become an increasingly significant issue.
1 code implementation • 7 Aug 2023 • Jen-tse Huang, Man Ho Lam, Eric John Li, Shujie Ren, Wenxuan Wang, Wenxiang Jiao, Zhaopeng Tu, Michael R. Lyu
Evaluating Large Language Models' (LLMs) anthropomorphic capabilities has become increasingly important in contemporary discourse.
1 code implementation • 14 Jun 2023 • Jianping Zhang, Zhuoer Xu, Shiwen Cui, Changhua Meng, Weibin Wu, Michael R. Lyu
Therefore, in this paper, we aim to analyze the robustness of latent diffusion models more thoroughly.
no code implementations • 8 Jun 2023 • Jinyang Liu, JunJie Huang, Yintong Huo, Zhihan Jiang, Jiazhen Gu, Zhuangbin Chen, Cong Feng, Minzhi Yan, Michael R. Lyu
System logs play a critical role in maintaining the reliability of software systems.
1 code implementation • 31 May 2023 • Jen-tse Huang, Wenxiang Jiao, Man Ho Lam, Eric John Li, Wenxuan Wang, Michael R. Lyu
Recent research has focused on examining Large Language Models' (LLMs) characteristics from a psychological standpoint, acknowledging the necessity of understanding their behavioral characteristics.
1 code implementation • CVPR 2023 • Jianping Zhang, Jen-tse Huang, Wenxuan Wang, Yichen Li, Weibin Wu, Xiaosen Wang, Yuxin Su, Michael R. Lyu
However, such methods selected the image augmentation path heuristically and may augment images that are semantics-inconsistent with the target images, which harms the transferability of the generated adversarial samples.
2 code implementations • CVPR 2023 • Jianping Zhang, Yizhan Huang, Weibin Wu, Michael R. Lyu
However, the variance of the back-propagated gradients in intermediate blocks of ViTs may still be large, which may make the generated adversarial samples focus on some model-specific features and get stuck in poor local optima.
2 code implementations • 14 Feb 2023 • Cheryl Lee, Tianyi Yang, Zhuangbin Chen, Yuxin Su, Yongqiang Yang, Michael R. Lyu
Our study demonstrates that logs and metrics can manifest system anomalies collaboratively and complementarily, and neither of them only is sufficient.
1 code implementation • 3 Nov 2022 • Yanyang Li, Jianqiao Zhao, Michael R. Lyu, LiWei Wang
Recent advances in large-scale pre-training provide large models with the potential to learn knowledge from the raw text.
1 code implementation • 24 Jul 2022 • Chaozheng Wang, Yuanhang Yang, Cuiyun Gao, Yun Peng, Hongyu Zhang, Michael R. Lyu
Besides, the performance of fine-tuning strongly relies on the amount of downstream data, while in practice, the scenarios with scarce data are common.
1 code implementation • Findings (NAACL) 2022 • Yifan Gao, Qingyu Yin, Zheng Li, Rui Meng, Tong Zhao, Bing Yin, Irwin King, Michael R. Lyu
Keyphrase generation is the task of automatically predicting keyphrases given a piece of long text.
no code implementations • 20 May 2022 • Wenxuan Wang, Wenxiang Jiao, Shuo Wang, Zhaopeng Tu, Michael R. Lyu
Zero-shot translation is a promising direction for building a comprehensive multilingual neural machine translation~(MNMT) system.
1 code implementation • 13 May 2022 • Jen-tse Huang, Jianping Zhang, Wenxuan Wang, Pinjia He, Yuxin Su, Michael R. Lyu
However, in practice, many of the generated test cases fail to preserve similar semantic meaning and are unnatural (e. g., grammar errors), which leads to a high false alarm rate and unnatural test cases.
1 code implementation • In2Writing (ACL) 2022 • Jingjing Li, Zichao Li, Tao Ge, Irwin King, Michael R. Lyu
In this approach, we simply fine-tune a pre-trained Transformer with masked language modeling and attribute classification.
2 code implementations • CVPR 2022 • Jianping Zhang, Weibin Wu, Jen-tse Huang, Yizhan Huang, Wenxuan Wang, Yuxin Su, Michael R. Lyu
Deep neural networks (DNNs) are known to be vulnerable to adversarial examples.
no code implementations • ACL 2022 • Wenchao Gu, Yanlin Wang, Lun Du, Hongyu Zhang, Shi Han, Dongmei Zhang, Michael R. Lyu
Code search is to search reusable code snippets from source code corpus based on natural languages queries.
no code implementations • 14 Feb 2022 • Jianqiao Zhao, Yanyang Li, Wanyu Du, Yangfeng Ji, Dong Yu, Michael R. Lyu, LiWei Wang
Hence, we propose segment act, an extension of dialog act from utterance level to segment level, and crowdsource a large-scale dataset for it.
no code implementations • 30 Sep 2021 • Haoli Bai, Lu Hou, Lifeng Shang, Xin Jiang, Irwin King, Michael R. Lyu
Experiments on GLUE and SQuAD benchmarks show that our proposed PTQ solution not only performs close to QAT, but also enjoys significant reductions in training time, memory overhead, and data consumption.
1 code implementation • 27 Aug 2021 • Zhuangbin Chen, Jinyang Liu, Yuxin Su, Hongyu Zhang, Xuemin Wen, Xiao Ling, Yongqiang Yang, Michael R. Lyu
The proposed framework is evaluated with real-world incident data collected from a large-scale online service system of Huawei Cloud.
1 code implementation • 13 Jul 2021 • Zhuangbin Chen, Jinyang Liu, Wenwei Gu, Yuxin Su, Michael R. Lyu
To better understand the characteristics of different anomaly detectors, in this paper, we provide a comprehensive review and evaluation of five popular neural networks used by six state-of-the-art methods.
1 code implementation • CVPR 2021 • Weibin Wu, Yuxin Su, Michael R. Lyu, Irwin King
Although deep neural networks (DNNs) have achieved tremendous performance in diverse vision challenges, they are surprisingly susceptible to adversarial examples, which are born of intentionally perturbing benign samples in a human-imperceptible fashion.
no code implementations • 8 Jun 2021 • Pengpeng Liu, Michael R. Lyu, Irwin King, Jia Xu
Then, a self-supervised learning framework is constructed: confident predictions from teacher models are served as annotations to guide the student model to learn optical flow for those less confident predictions.
1 code implementation • ACL 2021 • Wenxiang Jiao, Xing Wang, Zhaopeng Tu, Shuming Shi, Michael R. Lyu, Irwin King
In this work, we propose to improve the sampling procedure by selecting the most informative monolingual sentences to complement the parallel data.
no code implementations • 19 Apr 2021 • Shuzheng Gao, Cuiyun Gao, Yulan He, Jichuan Zeng, Lun Yiu Nie, Xin Xia, Michael R. Lyu
Code summaries help developers comprehend programs and reduce their time to infer the program functionalities during software maintenance.
2 code implementations • 17 Feb 2021 • Yifan Gao, Jingjing Li, Chien-Sheng Wu, Michael R. Lyu, Irwin King
On our created OR-ShARC dataset, MUDERN achieves the state-of-the-art performance, outperforming existing single-passage conversational machine reading models as well as a new multi-passage conversational machine reading baseline by a large margin.
1 code implementation • EMNLP 2020 • Yue Wang, Jing Li, Michael R. Lyu, Irwin King
Further analyses show that our multi-head attention is able to attend information from various aspects and boost classification or generation in diverse scenarios.
no code implementations • 10 Oct 2020 • Xixian Chen, Haiqin Yang, Shenglin Zhao, Michael R. Lyu, Irwin King
Estimating covariance matrix from massive high-dimensional and distributed data is significant for various real-world applications.
no code implementations • 10 Oct 2020 • Xixian Chen, Haiqin Yang, Shenglin Zhao, Michael R. Lyu, Irwin King
Data-dependent hashing methods have demonstrated good performance in various machine learning applications to learn a low-dimensional representation from the original data.
1 code implementation • EMNLP 2020 • Wenxiang Jiao, Xing Wang, Shilin He, Irwin King, Michael R. Lyu, Zhaopeng Tu
First, we train an identification model on the original training data, and use it to distinguish inactive examples and active examples by their sentence-level output probabilities.
1 code implementation • EMNLP 2020 • Yifan Gao, Chien-Sheng Wu, Jingjing Li, Shafiq Joty, Steven C. H. Hoi, Caiming Xiong, Irwin King, Michael R. Lyu
Based on the learned EDU and entailment representations, we either reply to the user our final decision "yes/no/irrelevant" of the initial question, or generate a follow-up question to inquiry more information.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yifan Gao, Piji Li, Wei Bi, Xiaojiang Liu, Michael R. Lyu, Irwin King
Besides a small number of high-resource sentence functions, a large portion of sentence functions is infrequent.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Wenxiang Jiao, Michael R. Lyu, Irwin King
Emotion Recognition in Conversations (ERC) aims to predict the emotional state of speakers in conversations, which is essentially a text classification task.
Emotion Recognition in Conversation General Classification +2
no code implementations • 23 Aug 2020 • Cuiyun Gao, Jichuan Zeng, Zhiyuan Wen, David Lo, Xin Xia, Irwin King, Michael R. Lyu
Experiments on popular apps from Google Play and Apple's App Store demonstrate the effectiveness of MERIT in identifying emerging app issues, improving the state-of-the-art method by 22. 3% in terms of F1-score.
8 code implementations • 14 Aug 2020 • Shilin He, Jieming Zhu, Pinjia He, Michael R. Lyu
To fill this significant gap and facilitate more research on AI-driven log analytics, we have collected and released loghub, a large collection of system log datasets.
Software Engineering
no code implementations • ACL 2020 • Jichuan Zeng, Xi Victoria Lin, Caiming Xiong, Richard Socher, Michael R. Lyu, Irwin King, Steven C. H. Hoi
Natural language interfaces to databases (NLIDB) democratize end user access to relational data.
no code implementations • NeurIPS 2020 • Jingjing Li, Zichao Li, Lili Mou, Xin Jiang, Michael R. Lyu, Irwin King
In this work, we present TGLS, a novel framework to unsupervised Text Generation by Learning from Search.
1 code implementation • 26 May 2020 • Yifan Gao, Chien-Sheng Wu, Shafiq Joty, Caiming Xiong, Richard Socher, Irwin King, Michael R. Lyu, Steven C. H. Hoi
The goal of conversational machine reading is to answer user questions given a knowledge base text which may require asking clarification questions.
1 code implementation • EMNLP 2020 • Yue Wang, Shafiq Joty, Michael R. Lyu, Irwin King, Caiming Xiong, Steven C. H. Hoi
By contrast, in this work, we propose VD-BERT, a simple yet effective framework of unified vision-dialog Transformer that leverages the pretrained BERT language models for Visual Dialog tasks.
no code implementations • 28 Apr 2020 • Shilin He, Xing Wang, Shuming Shi, Michael R. Lyu, Zhaopeng Tu
In this paper, we bridge the gap by assessing the bilingual knowledge learned by NMT models with phrase table -- an interpretable table of bilingual lexicons.
1 code implementation • 24 Apr 2020 • Bozhi Wu, Sen Chen, Cuiyun Gao, Lingling Fan, Yang Liu, Weiping Wen, Michael R. Lyu
In this paper, to fill this gap, we propose a novel and interpretable ML-based approach (named XMal) to classify malware with high accuracy and explain the classification result meanwhile.
no code implementations • 10 Feb 2020 • Jichuan Zeng, Jing Li, Yulan He, Cuiyun Gao, Michael R. Lyu, Irwin King
In our world with full of uncertainty, debates and argumentation contribute to the progress of science and society.
1 code implementation • 10 Feb 2020 • Cuiyun Gao, Jichuan Zeng, Xin Xia, David Lo, Michael R. Lyu, Irwin King
Previous studies showed that replying to a user review usually has a positive effect on the rating that is given by the user to the app.
no code implementations • 22 Nov 2019 • Jian Li, Xing Wang, Baosong Yang, Shuming Shi, Michael R. Lyu, Zhaopeng Tu
Starting from this intuition, we propose a novel approach to compose representations learned by different components in neural machine translation (e. g., multi-layer networks or multi-head attention), based on modeling strong interactions among neurons in the representation vectors.
1 code implementation • 20 Nov 2019 • Wenxiang Jiao, Michael R. Lyu, Irwin King
We propose an Attention Gated Hierarchical Memory Network (AGHMN) to address the problems of prior work: (1) Commonly used convolutional neural networks (CNNs) for utterance feature extraction are less compatible in the memory modules; (2) Unidirectional gated recurrent units (GRUs) only allow each historical utterance to have context before it, preventing information propagation in the opposite direction; (3) The Soft Attention for summarizing loses the positional and ordering information of memories, regardless of how the memory bank is built.
Ranked #50 on Emotion Recognition in Conversation on IEMOCAP
no code implementations • 21 Oct 2019 • Wenxiang Jiao, Irwin King, Michael R. Lyu
Word2Vec is the most popular model for word representation and has been widely investigated in literature.
1 code implementation • 20 Oct 2019 • Wenxiang Jiao, Michael R. Lyu, Irwin King
Witnessing the success of transfer learning in natural language process (NLP), we propose to pre-train a context-dependent encoder (CoDE) for ULER by learning from unlabeled conversation data.
no code implementations • IJCNLP 2019 • Jingjing Li, Yifan Gao, Lidong Bing, Irwin King, Michael R. Lyu
Question generation (QG) is the task of generating a question from a reference sentence and a specified answer within the sentence.
1 code implementation • 24 Sep 2019 • Jinyang Liu, Jieming Zhu, Shilin He, Pinjia He, Zibin Zheng, Michael R. Lyu
Data compression is essential to reduce the cost of log storage.
Databases Software Engineering
no code implementations • IJCNLP 2019 • Shilin He, Zhaopeng Tu, Xing Wang, Long-Yue Wang, Michael R. Lyu, Shuming Shi
Although neural machine translation (NMT) has advanced the state-of-the-art on various language pairs, the interpretability of NMT remains unsatisfactory.
no code implementations • WS 2019 • Fenglei Jin, Cuiyun Gao, Michael R. Lyu
In this paper, we propose a novel online topic tracking framework, named IEDL, for tracking the topic changes related to deep learning techniques on Stack Exchange and automatically interpreting each identified topic.
1 code implementation • ACL 2019 • Yifan Gao, Piji Li, Irwin King, Michael R. Lyu
The coreference alignment modeling explicitly aligns coreferent mentions in conversation history with corresponding pronominal references in generated questions, which makes generated questions interconnected to conversation history.
2 code implementations • ACL 2019 • Yue Wang, Jing Li, Hou Pong Chan, Irwin King, Michael R. Lyu, Shuming Shi
Further discussions show that our model learns meaningful topics, which interprets its superiority in social media keyphrase generation.
1 code implementation • 24 May 2019 • J. D. Curtó, I. C. Zarza, Kris Kitani, Irwin King, Michael R. Lyu
Dr. of Crosswise proposes a new architecture to reduce over-parametrization in Neural Networks.
1 code implementation • NAACL 2019 • Yue Wang, Jing Li, Irwin King, Michael R. Lyu, Shuming Shi
Automatic hashtag annotation plays an important role in content understanding for microblog posts.
1 code implementation • NAACL 2019 • Wenxiang Jiao, Haiqin Yang, Irwin King, Michael R. Lyu
In this paper, we address three challenges in utterance-level emotion recognition in dialogue systems: (1) the same word can deliver different emotions in different contexts; (2) some emotions are rarely seen in general dialogues; (3) long-range contextual information is hard to be effectively captured.
no code implementations • NAACL 2019 • Jian Li, Baosong Yang, Zi-Yi Dou, Xing Wang, Michael R. Lyu, Zhaopeng Tu
Multi-head attention is appealing for its ability to jointly extract different types of information from multiple representation subspaces.
1 code implementation • TACL 2019 • Jichuan Zeng, Jing Li, Yulan He, Cuiyun Gao, Michael R. Lyu, Irwin King
This paper presents an unsupervised framework for jointly modeling topic content and discourse behavior in microblog conversations.
1 code implementation • 25 Feb 2019 • Pengpeng Liu, Irwin King, Michael R. Lyu, Jia Xu
We present DDFlow, a data distillation approach to learning optical flow estimation from unlabeled data.
8 code implementations • 8 Nov 2018 • Jieming Zhu, Shilin He, Jinyang Liu, Pinjia He, Qi Xie, Zibin Zheng, Michael R. Lyu
Logs are imperative in the development and maintenance process of many software systems.
Software Engineering
no code implementations • NeurIPS 2018 • Han Shao, Xiaotian Yu, Irwin King, Michael R. Lyu
In this paper, under a weaker assumption on noises, we study the problem of \underline{lin}ear stochastic {\underline b}andits with h{\underline e}avy-{\underline t}ailed payoffs (LinBET), where the distributions have finite moments of order $1+\epsilon$, for some $\epsilon\in (0, 1]$.
no code implementations • EMNLP 2018 • Jian Li, Zhaopeng Tu, Baosong Yang, Michael R. Lyu, Tong Zhang
Multi-head attention is appealing for the ability to jointly attend to information from different representation subspaces at different positions.
no code implementations • EMNLP 2018 • Jichuan Zeng, Jing Li, Yan Song, Cuiyun Gao, Michael R. Lyu, Irwin King
Many classification models work poorly on short texts due to data sparsity.
2 code implementations • 8 Sep 2018 • Yifan Gao, Lidong Bing, Piji Li, Irwin King, Michael R. Lyu
We investigate the task of distractor generation for multiple choice reading comprehension questions from examinations.
no code implementations • 26 Aug 2018 • Wang Chen, Yifan Gao, Jiani Zhang, Irwin King, Michael R. Lyu
Keyphrase generation (KG) aims to generate a set of keyphrases given a document, which is a fundamental task in natural language processing (NLP).
no code implementations • 10 Jul 2018 • Yifan Gao, Lidong Bing, Wang Chen, Michael R. Lyu, Irwin King
We investigate the difficulty levels of questions in reading comprehension datasets such as SQuAD, and propose a new question generation setting, named Difficulty-controllable Question Generation (DQG).
no code implementations • 27 Jun 2018 • Hui Xu, Yuxin Su, Zirui Zhao, Yangfan Zhou, Michael R. Lyu, Irwin King
Our obfuscation approach is very effective to protect the critical structure of a deep learning model from being exposed to attackers.
Cryptography and Security
no code implementations • 12 Jun 2018 • Pinjia He, Jieming Zhu, Pengcheng Xu, Zibin Zheng, Michael R. Lyu
A typical log-based system reliability management procedure is to first parse log messages because of their unstructured format; and apply data mining techniques on the parsed logs to obtain critical system behavior information.
Software Engineering
1 code implementation • 27 Nov 2017 • Jian Li, Yue Wang, Michael R. Lyu, Irwin King
Intelligent code completion has become an essential research task to accelerate modern software development.
no code implementations • 26 Nov 2017 • Pengpeng Liu, Xiaojuan Qi, Pinjia He, Yikang Li, Michael R. Lyu, Irwin King
Image completion has achieved significant progress due to advances in generative adversarial networks (GANs).
1 code implementation • 17 Nov 2017 • Joachim D. Curtó, Irene C. Zarza, Fernando de la Torre, Irwin King, Michael R. Lyu
Generative Adversarial Networks (GANs) convergence in a high-resolution setting with a computational constrain of GPU memory capacity (from 12GB to 24 GB) has been beset with difficulty due to the known lack of convergence rate stability.
Ranked #1 on Image Generation on CelebA 128x128 (MS-SSIM metric)
no code implementations • 3 Oct 2017 • Hui Xu, Yangfan Zhou, Yu Kang, Michael R. Lyu
On the other hand, the performance requirement for model-oriented obfuscation approaches is too weak to develop practical program obfuscation solutions.
Cryptography and Security Software Engineering
no code implementations • ICML 2017 • Xixian Chen, Michael R. Lyu, Irwin King
Estimating covariance matrices is a fundamental technique in various domains, most notably in machine learning and signal processing.
no code implementations • 3 Jul 2016 • Shenglin Zhao, Irwin King, Michael R. Lyu
Then, we present a comprehensive review in three aspects: influential factors for POI recommendation, methodologies employed for POI recommendation, and different tasks in POI recommendation.
no code implementations • 8 Jun 2015 • Hui Xu, Yangfan Zhou, Michael R. Lyu
Our idea is to impede the replication of tampering via program diversification, and thus increasing the complexity to break the whole software system.
Cryptography and Security Programming Languages
no code implementations • NeurIPS 2014 • Shouyuan Chen, Tian Lin, Irwin King, Michael R. Lyu, Wei Chen
We also establish a general problem-dependent lower bound for the CPE problem.
no code implementations • NeurIPS 2013 • Shouyuan Chen, Michael R. Lyu, Irwin King, Zenglin Xu
For the noisy cases, we also prove error bounds for a constrained convex program for recovering the tensors.