Search Results for author: Ben Athiwaratkun

Found 24 papers, 12 papers with code

Improving Model Alignment Through Collective Intelligence of Open-Source LLMS

no code implementations5 May 2025 Junlin Wang, Roy Xie, Shang Zhu, Jue Wang, Ben Athiwaratkun, Bhuwan Dhingra, Shuaiwen Leon Song, Ce Zhang, James Zou

Building helpful and harmless large language models (LLMs) requires effective model alignment approach based on human instructions and feedback, which necessitates high-quality human-labeled data.

How Well Can General Vision-Language Models Learn Medicine By Watching Public Educational Videos?

no code implementations19 Apr 2025 Rahul Thapa, Andrew Li, Qingyang Wu, Bryan He, Yuki Sahashi, Christina Binder, Angela Zhang, Ben Athiwaratkun, Shuaiwen Leon Song, David Ouyang, James Zou

On these benchmarks, the 2B model achieves gains of 99. 1% and 98. 1%, while the 7B model shows gains of 22. 5% and 52. 1%, respectively, demonstrating the models' ability to generalize and perform biomedical video understanding on cleaner and more standardized datasets than those seen during training.

Video Understanding

Think Deep, Think Fast: Investigating Efficiency of Verifier-free Inference-time-scaling Methods

no code implementations18 Apr 2025 Junlin Wang, Shang Zhu, Jon Saad-Falcon, Ben Athiwaratkun, Qingyang Wu, Jue Wang, Shuaiwen Leon Song, Ce Zhang, Bhuwan Dhingra, James Zou

There is intense interest in investigating how inference time compute (ITC) (e. g. repeated sampling, refinements, etc) can improve large language model (LLM) capabilities.

Large Language Model

Scaling Instruction-Tuned LLMs to Million-Token Contexts via Hierarchical Synthetic Data Generation

no code implementations17 Apr 2025 Linda He, Jue Wang, Maurice Weber, Shang Zhu, Ben Athiwaratkun, Ce Zhang

Large Language Models (LLMs) struggle with long-context reasoning, not only due to the quadratic scaling of computational complexity with sequence length but also because of the scarcity and expense of annotating long-context data.

Synthetic Data Generation

Training-Free Activation Sparsity in Large Language Models

1 code implementation26 Aug 2024 James Liu, Pragaash Ponnusamy, Tianle Cai, Han Guo, Yoon Kim, Ben Athiwaratkun

Activation sparsity can enable practical inference speedups in large language models (LLMs) by reducing the compute and memory-movement required for matrix multiplications during the forward pass.

Quantization

Reasoning in Token Economies: Budget-Aware Evaluation of LLM Reasoning Strategies

no code implementations10 Jun 2024 Junlin Wang, Siddhartha Jain, Dejiao Zhang, Baishakhi Ray, Varun Kumar, Ben Athiwaratkun

A diverse array of reasoning strategies has been proposed to elicit the capabilities of large language models.

Ingenuity

Mixture-of-Agents Enhances Large Language Model Capabilities

3 code implementations7 Jun 2024 Junlin Wang, Jue Wang, Ben Athiwaratkun, Ce Zhang, James Zou

With the growing number of LLMs, how to harness the collective expertise of multiple LLMs is an exciting open direction.

Language Modeling Language Modelling +2

Dragonfly: Multi-Resolution Zoom-In Encoding Enhances Vision-Language Models

1 code implementation3 Jun 2024 Rahul Thapa, Kezhen Chen, Ian Covert, Rahul Chalamala, Ben Athiwaratkun, Shuaiwen Leon Song, James Zou

Recent advances in vision-language models (VLMs) have demonstrated the advantages of processing images at higher resolutions and utilizing multi-crop features to preserve native resolution details.

Image Captioning Language Modelling +3

Multi-lingual Evaluation of Code Generation Models

2 code implementations26 Oct 2022 Ben Athiwaratkun, Sanjay Krishna Gouda, Zijian Wang, Xiaopeng Li, Yuchen Tian, Ming Tan, Wasi Uddin Ahmad, Shiqi Wang, Qing Sun, Mingyue Shang, Sujan Kumar Gonugondla, Hantian Ding, Varun Kumar, Nathan Fulton, Arash Farahani, Siddhartha Jain, Robert Giaquinto, Haifeng Qian, Murali Krishna Ramanathan, Ramesh Nallapati, Baishakhi Ray, Parminder Bhatia, Sudipta Sengupta, Dan Roth, Bing Xiang

Using these benchmarks, we are able to assess the performance of code generation models in a multi-lingual fashion, and discovered generalization ability of language models on out-of-domain languages, advantages of multi-lingual models over mono-lingual, the ability of few-shot prompting to teach the model new languages, and zero-shot translation abilities even on mono-lingual settings.

Code Completion Code Translation +2

Joint Text and Label Generation for Spoken Language Understanding

no code implementations11 May 2021 Yang Li, Ben Athiwaratkun, Cicero Nogueira dos santos, Bing Xiang

In this work, we propose to leverage the prior information embedded in pretrained language models (LM) to improve generalization for intent classification and slot labeling tasks with limited training data.

intent-classification Intent Classification +2

Structured Prediction as Translation between Augmented Natural Languages

2 code implementations ICLR 2021 Giovanni Paolini, Ben Athiwaratkun, Jason Krone, Jie Ma, Alessandro Achille, Rishita Anubhai, Cicero Nogueira dos santos, Bing Xiang, Stefano Soatto

We propose a new framework, Translation between Augmented Natural Languages (TANL), to solve many structured prediction language tasks including joint entity and relation extraction, nested named entity recognition, relation classification, semantic role labeling, event extraction, coreference resolution, and dialogue state tracking.

coreference-resolution Dialogue State Tracking +12

Augmented Natural Language for Generative Sequence Labeling

no code implementations EMNLP 2020 Ben Athiwaratkun, Cicero Nogueira dos santos, Jason Krone, Bing Xiang

We set a new state-of-the-art for few-shot slot labeling, improving substantially upon the previous 5-shot ($75. 0\% \rightarrow 90. 9\%$) and 1-shot ($70. 4\% \rightarrow 81. 0\%$) state-of-the-art results.

intent-classification Intent Classification +4

There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average

2 code implementations ICLR 2019 Ben Athiwaratkun, Marc Finzi, Pavel Izmailov, Andrew Gordon Wilson

Presently the most successful approaches to semi-supervised learning are based on consistency regularization, whereby a model is trained to be robust to small perturbations of its inputs and parameters.

Domain Adaptation Semi-Supervised Image Classification

Probabilistic FastText for Multi-Sense Word Embeddings

1 code implementation ACL 2018 Ben Athiwaratkun, Andrew Gordon Wilson, Anima Anandkumar

We introduce Probabilistic FastText, a new model for word embeddings that can capture multiple word senses, sub-word structure, and uncertainty information.

Word Embeddings Word Similarity

Hierarchical Density Order Embeddings

2 code implementations ICLR 2018 Ben Athiwaratkun, Andrew Gordon Wilson

By representing words with probability densities rather than point vectors, probabilistic word embeddings can capture rich and interpretable semantic information and uncertainty.

Lexical Entailment Word Embeddings

Multimodal Word Distributions

2 code implementations ACL 2017 Ben Athiwaratkun, Andrew Gordon Wilson

Word embeddings provide point representations of words containing useful semantic information.

Word Embeddings Word Similarity

Adversarial Deep Averaging Networks for Cross-Lingual Sentiment Classification

2 code implementations TACL 2018 Xilun Chen, Yu Sun, Ben Athiwaratkun, Claire Cardie, Kilian Weinberger

To tackle the sentiment classification problem in low-resource languages without adequate annotated data, we propose an Adversarial Deep Averaging Network (ADAN) to transfer the knowledge learned from labeled data on a resource-rich source language to low-resource languages where only unlabeled data exists.

Classification Cross-Lingual Document Classification +5

Feature Representation in Convolutional Neural Networks

no code implementations8 Jul 2015 Ben Athiwaratkun, Keegan Kang

Our results show that CNN feature maps can be used with Random Forests and SVM to yield classification results that outperforms the original CNN.

Classification General Classification +1

Cannot find the paper you are looking for? You can Submit a new open access paper.