Search Results for author: Davis Wertheimer

Found 8 papers, 4 papers with code

INDUS: Effective and Efficient Language Models for Scientific Applications

no code implementations17 May 2024 Bishwaranjan Bhattacharjee, Aashka Trivedi, Masayasu Muraoka, Muthukumaran Ramasubramanian, Takuma Udagawa, Iksha Gurung, Nishan Pantha, Rong Zhang, Bharath Dandala, Rahul Ramachandran, Manil Maskey, Kaylin Bugbee, Mike Little, Elizabeth Fancher, Irina Gerasimov, Armin Mehrabian, Lauren Sanders, Sylvain Costes, Sergi Blanco-Cuaresma, Kelly Lockhart, Thomas Allen, Felix Grezes, Megan Ansdell, Alberto Accomazzi, Yousef El-Kurdi, Davis Wertheimer, Birgit Pfitzmann, Cesar Berrospi Ramis, Michele Dolfi, Rafael Teixeira de Lima, Panagiotis Vagenas, S. Karthik Mukkavilli, Peter Staar, Sanaz Vahidinia, Ryan McGranaghan, Tsendgar Lee

The suite of models include: (1) an encoder model trained using domain-specific vocabulary and corpora to address NLP tasks, (2) a contrastive-learning based text embedding model trained using a diverse set of datasets to address information retrieval tasks and (3) smaller versions of these models created using knowledge distillation for applications which have latency or resource constraints.

Contrastive Learning Information Retrieval +4

Accelerating Production LLMs with Combined Token/Embedding Speculators

1 code implementation29 Apr 2024 Davis Wertheimer, Joshua Rosenkranz, Thomas Parnell, Sahil Suneja, Pavithra Ranganathan, Raghu Ganti, Mudhakar Srivatsa

This technical report describes the design and training of novel speculative decoding draft models, for accelerating the inference speeds of large language models in a production environment.

SudokuSens: Enhancing Deep Learning Robustness for IoT Sensing Applications using a Generative Approach

no code implementations3 Feb 2024 Tianshi Wang, Jinyang Li, Ruijie Wang, Denizhan Kara, Shengzhong Liu, Davis Wertheimer, Antoni Viros-i-Martin, Raghu Ganti, Mudhakar Srivatsa, Tarek Abdelzaher

To incorporate sufficient diversity into the IoT training data, one therefore needs to consider a combinatorial explosion of training cases that are multiplicative in the number of objects considered and the possible environmental conditions in which such objects may be encountered.

Contrastive Learning

Diagnosing and Remedying Shot Sensitivity with Cosine Few-Shot Learners

no code implementations7 Jul 2022 Davis Wertheimer, Luming Tang, Bharath Hariharan

Existing approaches generally assume that the shot number at test time is known in advance.

Novel Concepts

Revisiting Pose-Normalization for Fine-Grained Few-Shot Recognition

1 code implementation CVPR 2020 Luming Tang, Davis Wertheimer, Bharath Hariharan

Few-shot, fine-grained classification requires a model to learn subtle, fine-grained distinctions between different classes (e. g., birds) based on a few images alone.

Classification General Classification

Few-Shot Learning with Localization in Realistic Settings

1 code implementation CVPR 2019 Davis Wertheimer, Bharath Hariharan

Traditional recognition methods typically require large, artificially-balanced training classes, while few-shot learning methods are tested on artificially small ones.

Few-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.