Search Results for author: Jared Fernandez

Found 8 papers, 4 papers with code

Holistically Evaluating the Environmental Impact of Creating Language Models

no code implementations3 Mar 2025 Jacob Morrison, Clara Na, Jared Fernandez, Tim Dettmers, Emma Strubell, Jesse Dodge

As the performance of artificial intelligence systems has dramatically increased, so too has the environmental impact of creating these systems.

Hardware Scaling Trends and Diminishing Returns in Large-Scale Distributed Training

no code implementations20 Nov 2024 Jared Fernandez, Luca Wehrstedt, Leonid Shamis, Mostafa Elhoushi, Kalyan Saladi, Yonatan Bisk, Emma Strubell, Jacob Kahn

Dramatic increases in the capabilities of neural network models in recent years are driven by scaling model size, training data, and corresponding computational resources.

Gradient Localization Improves Lifelong Pretraining of Language Models

no code implementations7 Nov 2024 Jared Fernandez, Yonatan Bisk, Emma Strubell

Large Language Models (LLMs) trained on web-scale text corpora have been shown to capture world knowledge in their parameters.

Continual Learning World Knowledge

The Framework Tax: Disparities Between Inference Efficiency in NLP Research and Deployment

1 code implementation13 Feb 2023 Jared Fernandez, Jacob Kahn, Clara Na, Yonatan Bisk, Emma Strubell

In this work, we examine this phenomenon through a series of case studies analyzing the effects of model design decisions, framework paradigms, and hardware platforms on total model latency.

Computational Efficiency

CIGLI: Conditional Image Generation from Language & Image

1 code implementation20 Aug 2021 Xiaopeng Lu, Lynnette Ng, Jared Fernandez, Hao Zhu

Instead of generating an image based on text as in text-image generation, this task requires the generation of an image from a textual description and an image prompt.

Conditional Image Generation

CODAH: An Adversarially Authored Question-Answer Dataset for Common Sense

2 code implementations8 Apr 2019 Michael Chen, Mike D'Arcy, Alisa Liu, Jared Fernandez, Doug Downey

To produce a more difficult dataset, we introduce a novel procedure for question acquisition in which workers author questions designed to target weaknesses of state-of-the-art neural question answering systems.

 Ranked #1 on Common Sense Reasoning on CODAH (using extra training data)

Common Sense Reasoning Question Answering +2

Cannot find the paper you are looking for? You can Submit a new open access paper.