no code implementations • 17 May 2024 • Bishwaranjan Bhattacharjee, Aashka Trivedi, Masayasu Muraoka, Muthukumaran Ramasubramanian, Takuma Udagawa, Iksha Gurung, Nishan Pantha, Rong Zhang, Bharath Dandala, Rahul Ramachandran, Manil Maskey, Kaylin Bugbee, Mike Little, Elizabeth Fancher, Irina Gerasimov, Armin Mehrabian, Lauren Sanders, Sylvain Costes, Sergi Blanco-Cuaresma, Kelly Lockhart, Thomas Allen, Felix Grezes, Megan Ansdell, Alberto Accomazzi, Yousef El-Kurdi, Davis Wertheimer, Birgit Pfitzmann, Cesar Berrospi Ramis, Michele Dolfi, Rafael Teixeira de Lima, Panagiotis Vagenas, S. Karthik Mukkavilli, Peter Staar, Sanaz Vahidinia, Ryan McGranaghan, Tsendgar Lee
The suite of models include: (1) an encoder model trained using domain-specific vocabulary and corpora to address NLP tasks, (2) a contrastive-learning based text embedding model trained using a diverse set of datasets to address information retrieval tasks and (3) smaller versions of these models created using knowledge distillation for applications which have latency or resource constraints.
1 code implementation • 29 Apr 2024 • Davis Wertheimer, Joshua Rosenkranz, Thomas Parnell, Sahil Suneja, Pavithra Ranganathan, Raghu Ganti, Mudhakar Srivatsa
This technical report describes the design and training of novel speculative decoding draft models, for accelerating the inference speeds of large language models in a production environment.
no code implementations • 3 Feb 2024 • Tianshi Wang, Jinyang Li, Ruijie Wang, Denizhan Kara, Shengzhong Liu, Davis Wertheimer, Antoni Viros-i-Martin, Raghu Ganti, Mudhakar Srivatsa, Tarek Abdelzaher
To incorporate sufficient diversity into the IoT training data, one therefore needs to consider a combinatorial explosion of training cases that are multiplicative in the number of objects considered and the possible environmental conditions in which such objects may be encountered.
no code implementations • 7 Jul 2022 • Davis Wertheimer, Luming Tang, Bharath Hariharan
Existing approaches generally assume that the shot number at test time is known in advance.
1 code implementation • CVPR 2021 • Davis Wertheimer, Luming Tang, Bharath Hariharan
In this paper we reformulate few-shot classification as a reconstruction problem in latent space.
no code implementations • 25 Nov 2020 • Davis Wertheimer, Omid Poursaeed, Bharath Hariharan
We aim to build image generation models that generalize to new domains from few examples.
1 code implementation • CVPR 2020 • Luming Tang, Davis Wertheimer, Bharath Hariharan
Few-shot, fine-grained classification requires a model to learn subtle, fine-grained distinctions between different classes (e. g., birds) based on a few images alone.
1 code implementation • CVPR 2019 • Davis Wertheimer, Bharath Hariharan
Traditional recognition methods typically require large, artificially-balanced training classes, while few-shot learning methods are tested on artificially small ones.