Search Results for author: Aditya G. Parameswaran

Found 4 papers, 0 papers with code

Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences

no code implementations18 Apr 2024 Shreya Shankar, J. D. Zamfirescu-Pereira, Björn Hartmann, Aditya G. Parameswaran, Ian Arawjo

In particular, we identify a phenomenon we dub \emph{criteria drift}: users need criteria to grade outputs, but grading outputs helps users define criteria.

Revisiting Prompt Engineering via Declarative Crowdsourcing

no code implementations7 Aug 2023 Aditya G. Parameswaran, Shreya Shankar, Parth Asawa, Naman jain, Yujie Wang

Large language models (LLMs) are incredibly powerful at comprehending and generating data in the form of text, but are brittle and error-prone.

Entity Resolution Imputation +1

Operationalizing Machine Learning: An Interview Study

no code implementations16 Sep 2022 Shreya Shankar, Rolando Garcia, Joseph M. Hellerstein, Aditya G. Parameswaran

Organizations rely on machine learning engineers (MLEs) to operationalize ML, i. e., deploy and maintain ML pipelines in production.

Autonomous Vehicles

Rethinking Streaming Machine Learning Evaluation

no code implementations23 May 2022 Shreya Shankar, Bernease Herman, Aditya G. Parameswaran

While most work on evaluating machine learning (ML) models focuses on computing accuracy on batches of data, tracking accuracy alone in a streaming setting (i. e., unbounded, timestamp-ordered datasets) fails to appropriately identify when models are performing unexpectedly.

BIG-bench Machine Learning Position

Cannot find the paper you are looking for? You can Submit a new open access paper.