no code implementations • 28 May 2024 • Nan Jiang, Xiaopeng Li, Shiqi Wang, Qiang Zhou, Soneya Binta Hossain, Baishakhi Ray, Varun Kumar, Xiaofei Ma, Anoop Deoras
We thus propose an automated pipeline to collect a high-quality dataset for code explanation and refinement by generating a number of explanations and refinement trajectories and filtering via execution verification.
1 code implementation • 22 May 2024 • Gauthier Guinet, Behrooz Omidvar-Tehrani, Anoop Deoras, Laurent Callot
Evaluation is performed by scoring the RAG on an automatically-generated synthetic exam composed of multiple choice questions based on the corpus of documents associated with the task.
no code implementations • 6 May 2024 • Tao Yu, Gaurav Gupta, Karthick Gopalswamy, Amith Mamidala, Hao Zhou, Jeffrey Huynh, Youngsuk Park, Ron Diamant, Anoop Deoras, Luke Huan
Large models training is plagued by the intense compute cost and limited hardware memory.
no code implementations • 24 Apr 2024 • Haifeng Qian, Sujan Kumar Gonugondla, Sungsoo Ha, Mingyue Shang, Sanjay Krishna Gouda, Ramesh Nallapati, Sudipta Sengupta, Xiaofei Ma, Anoop Deoras
Speculative decoding has emerged as a powerful method to improve latency and throughput in hosting large language models.
no code implementations • 16 Apr 2024 • Hantian Ding, Zijian Wang, Giovanni Paolini, Varun Kumar, Anoop Deoras, Dan Roth, Stefano Soatto
In large language model training, input documents are typically concatenated together and then split into sequences of equal length to avoid padding tokens.
no code implementations • 11 Apr 2024 • Yuhao Zhang, Shiqi Wang, Haifeng Qian, Zijian Wang, Mingyue Shang, Linbo Liu, Sanjay Krishna Gouda, Baishakhi Ray, Murali Krishna Ramanathan, Xiaofei Ma, Anoop Deoras
Code generation models are not robust to small perturbations, which often lead to inconsistent and incorrect generations and significantly degrade the performance of these models.
no code implementations • 22 Dec 2023 • Behnam Rahdari, Hao Ding, Ziwei Fan, Yifei Ma, Zhuotong Chen, Anoop Deoras, Branislav Kveton
The unique capabilities of Large Language Models (LLMs), such as the natural language text generation ability, position them as strong candidates for providing explanation for recommendations.
1 code implementation • 30 Oct 2023 • Ziqian Lin, Hao Ding, Nghia Trong Hoang, Branislav Kveton, Anoop Deoras, Hao Wang
In particular, we propose to develop a generic recommender that captures universal interaction patterns by training on generic user-item interaction data extracted from different domains, which can then be fast adapted to improve few-shot learning performance in unseen new domains (with limited data).
no code implementations • 11 Jul 2023 • Siddhartha Jain, Xiaofei Ma, Anoop Deoras, Bing Xiang
We show strong improvements for selecting the best k generations for code generation tasks as well as robust improvements for the best generation for the tasks of autoformalization, summarization, and translation.
no code implementations • 13 Jun 2023 • Anusha Lalitha, Kousha Kalantari, Yifei Ma, Anoop Deoras, Branislav Kveton
Our algorithms rely on non-uniform budget allocations among the arms where the arms with higher reward variances are pulled more often than those with lower variances.
no code implementations • 5 Jun 2023 • Ziwei Fan, Hao Ding, Anoop Deoras, Trong Nghia Hoang
To mitigate this data bottleneck, we postulate that recommendation patterns learned from existing mature market segments (with private data) could be adapted to build effective warm-start models for emerging ones.
no code implementations • 31 May 2022 • Mostafa Rahmani, Anoop Deoras, Laurent Callot
This paper presents a novel, closed-form, and data/computation efficient online anomaly detection algorithm for time-series data.
no code implementations • 18 Mar 2022 • Trong Nghia Hoang, Anoop Deoras, Tong Zhao, Jin Li, George Karypis
We develop and investigate a personalizable deep metric model that captures both the internal contents of items and how they were interacted with by users.
no code implementations • ICLR 2022 • Yifei Ma, Ge Liu, Anoop Deoras
RIM allows us to rethink recommendation in a Matching (Mtch) scenario, where the benefits of the users (e. g., ItemRec relevance) and item providers (e. g., item-exposure guarantees) are considered at the same time.
no code implementations • NeurIPS Workshop ICBINB 2021 • Yuhui Zhang, Hao Ding, Zeren Shui, Yifei Ma, James Zou, Anoop Deoras, Hao Wang
Pre-trained language models (PLMs) such as BERT and GPT learn general text representations and encode extensive world knowledge; thus, they can be efficiently and accurately adapted to various downstream tasks.
no code implementations • 18 May 2021 • Hao Ding, Yifei Ma, Anoop Deoras, Yuyang Wang, Hao Wang
This poses a chicken-and-egg problem for early-stage products, whose amount of data, in turn, relies on the performance of their RS.