Search Results for author: Raymond Li

Found 20 papers, 11 papers with code

T3-Vis: visual analytic for Training and fine-Tuning Transformers in NLP

1 code implementation EMNLP (ACL) 2021 Raymond Li, Wen Xiao, Lanjun Wang, Hyeju Jang, Giuseppe Carenini

Transformers are the dominant architecture in NLP, but their training and fine-tuning is still very challenging.

SQL-Encoder: Improving NL2SQL In-Context Learning Through a Context-Aware Encoder

no code implementations24 Mar 2024 Mohammadreza Pourreza, Davood Rafiei, Yuxi Feng, Raymond Li, Zhenan Fan, Weiwei Zhang

Furthermore, compared to these competitive models, our proposed encoder enhances the downstream performance of NL2SQL models in 1-shot in-context learning scenarios by 1-2\% for GPT-3. 5-turbo, 4-8\% for CodeLlama-7B, and 2-3\% for CodeLlama-13B.

In-Context Learning

Visual Analytics for Generative Transformer Models

no code implementations21 Nov 2023 Raymond Li, Ruixin Yang, Wen Xiao, Ahmed Aburaed, Gabriel Murray, Giuseppe Carenini

While transformer-based models have achieved state-of-the-art results in a variety of classification and generation tasks, their black-box nature makes them challenging for interpretability.

Mixture-of-Linguistic-Experts Adapters for Improving and Interpreting Pre-trained Language Models

no code implementations24 Oct 2023 Raymond Li, Gabriel Murray, Giuseppe Carenini

In this work, we propose a method that combines two popular research areas by injecting linguistic structures into pre-trained language models in the parameter-efficient fine-tuning (PEFT) setting.

Diversity-Aware Coherence Loss for Improving Neural Topic Models

1 code implementation25 May 2023 Raymond Li, Felipe González-Pizarro, Linzi Xing, Gabriel Murray, Giuseppe Carenini

The standard approach for neural topic modeling uses a variational autoencoder (VAE) framework that jointly minimizes the KL divergence between the estimated posterior and prior, in addition to the reconstruction loss.

Topic Models

StarCoder: may the source be with you!

4 code implementations9 May 2023 Raymond Li, Loubna Ben allal, Yangtian Zi, Niklas Muennighoff, Denis Kocetkov, Chenghao Mou, Marc Marone, Christopher Akiki, Jia Li, Jenny Chim, Qian Liu, Evgenii Zheltonozhskii, Terry Yue Zhuo, Thomas Wang, Olivier Dehaene, Mishig Davaadorj, Joel Lamy-Poirier, João Monteiro, Oleh Shliazhko, Nicolas Gontier, Nicholas Meade, Armel Zebaze, Ming-Ho Yee, Logesh Kumar Umapathi, Jian Zhu, Benjamin Lipkin, Muhtasham Oblokulov, Zhiruo Wang, Rudra Murthy, Jason Stillerman, Siva Sankalp Patel, Dmitry Abulkhanov, Marco Zocca, Manan Dey, Zhihan Zhang, Nour Fahmy, Urvashi Bhattacharyya, Wenhao Yu, Swayam Singh, Sasha Luccioni, Paulo Villegas, Maxim Kunakov, Fedor Zhdanov, Manuel Romero, Tony Lee, Nadav Timor, Jennifer Ding, Claire Schlesinger, Hailey Schoelkopf, Jan Ebert, Tri Dao, Mayank Mishra, Alex Gu, Jennifer Robinson, Carolyn Jane Anderson, Brendan Dolan-Gavitt, Danish Contractor, Siva Reddy, Daniel Fried, Dzmitry Bahdanau, Yacine Jernite, Carlos Muñoz Ferrandis, Sean Hughes, Thomas Wolf, Arjun Guha, Leandro von Werra, Harm de Vries

The BigCode community, an open-scientific collaboration working on the responsible development of Large Language Models for Code (Code LLMs), introduces StarCoder and StarCoderBase: 15. 5B parameter models with 8K context length, infilling capabilities and fast large-batch inference enabled by multi-query attention.

8k Code Generation

NL4Opt Competition: Formulating Optimization Problems Based on Their Natural Language Descriptions

1 code implementation14 Mar 2023 Rindranirina Ramamonjison, Timothy T. Yu, Raymond Li, Haley Li, Giuseppe Carenini, Bissan Ghaddar, Shiqi He, Mahdi Mostajabdaveh, Amin Banitalebi-Dehkordi, Zirui Zhou, Yong Zhang

The Natural Language for Optimization (NL4Opt) Competition was created to investigate methods of extracting the meaning and formulation of an optimization problem based on its text description.

Language Modelling Large Language Model

The Stack: 3 TB of permissively licensed source code

no code implementations20 Nov 2022 Denis Kocetkov, Raymond Li, Loubna Ben allal, Jia Li, Chenghao Mou, Carlos Muñoz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, Harm de Vries

Large Language Models (LLMs) play an ever-increasing role in the field of Artificial Intelligence (AI)--not only for natural language processing but also for code understanding and generation.

T3-Vis: a visual analytic framework for Training and fine-Tuning Transformers in NLP

1 code implementation31 Aug 2021 Raymond Li, Wen Xiao, Lanjun Wang, Hyeju Jang, Giuseppe Carenini

Transformers are the dominant architecture in NLP, but their training and fine-tuning is still very challenging.

ConVIScope: Visual Analytics for Exploring Patient Conversations

no code implementations30 Aug 2021 Raymond Li, Enamul Hoque, Giuseppe Carenini, Richard Lester, Raymond Chau

The proliferation of text messaging for mobile health is generating a large amount of patient-doctor conversations that can be extremely valuable to health care professionals.

byteSteady: Fast Classification Using Byte-Level n-Gram Embeddings

no code implementations24 Jun 2021 Xiang Zhang, Alexandre Drouin, Raymond Li

This article introduces byteSteady -- a fast model for classification using byte-level n-gram embeddings.

text-classification Text Classification

DuoRAT: Towards Simpler Text-to-SQL Models

1 code implementation NAACL 2021 Torsten Scholak, Raymond Li, Dzmitry Bahdanau, Harm de Vries, Chris Pal

Recent neural text-to-SQL models can effectively translate natural language questions to corresponding SQL queries on unseen databases.

Text-To-SQL

Video Monitoring Queries

no code implementations24 Feb 2020 Nick Koudas, Raymond Li, Ioannis Xarchakos

We demonstrate that the application of the techniques proposed in conjunction with declarative queries on video streams can dramatically increase the frame processing rate and speed up query processing by at least two orders of magnitude.

Image Classification object-detection +1

Cannot find the paper you are looking for? You can Submit a new open access paper.