Search Results for author: Chris Wendler

Found 16 papers, 12 papers with code

Controlling Latent Diffusion Using Latent CLIP

1 code implementation11 Mar 2025 Jason Becker, Chris Wendler, Peter Baylies, Robert West, Christian Wressnegger

We train Latent-CLIP on 2. 7B pairs of latent images and descriptive texts, and show that it matches zero-shot classification performance of similarly sized CLIP models on both the ImageNet benchmark and a LDM-generated version of it, demonstrating its effectiveness in assessing both real and generated content.

Denoising Descriptive +1

Byte BPE Tokenization as an Inverse string Homomorphism

no code implementations4 Dec 2024 Saibo Geng, Sankalp Gambhir, Chris Wendler, Robert West

Tokenization is an important preprocessing step in the training and inference of large language models (LLMs).

Controllable Context Sensitivity and the Knob Behind It

1 code implementation11 Nov 2024 Julian Minder, Kevin Du, Niklas Stoehr, Giovanni Monea, Chris Wendler, Robert West, Ryan Cotterell

In this paper, we search for a knob which controls this sensitivity, determining whether language models answer from the context or their prior knowledge.

Question Answering

Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders

1 code implementation28 Oct 2024 Viacheslav Surkov, Chris Wendler, Mikhail Terekhov, Justin Deschenaux, Robert West, Caglar Gulcehre

We investigated the possibility of using SAEs to learn interpretable features for a few-step text-to-image diffusion models, such as SDXL Turbo.

Denoising

Multilingual Contrastive Decoding via Language-Agnostic Layers Skipping

1 code implementation15 Jul 2024 Wenhao Zhu, Sizhe Liu, ShuJian Huang, Shuaijie She, Chris Wendler, Jiajun Chen

Decoding by contrasting layers (DoLa), is designed to improve the generation quality of large language models (LLMs) by contrasting the prediction probabilities between an early exit output (amateur logits) and the final output (expert logits).

Do Llamas Work in English? On the Latent Language of Multilingual Transformers

1 code implementation16 Feb 2024 Chris Wendler, Veniamin Veselovsky, Giovanni Monea, Robert West

Tracking intermediate embeddings through their high-dimensional space reveals three distinct phases, whereby intermediate embeddings (1) start far away from output token embeddings; (2) already allow for decoding a semantically correct next token in the middle layers, but give higher probability to its version in English than in the input language; (3) finally move into an input-language-specific region of the embedding space.

Sketch-Guided Constrained Decoding for Boosting Blackbox Large Language Models without Logit Access

1 code implementation18 Jan 2024 Saibo Geng, Berkay Döner, Chris Wendler, Martin Josifoski, Robert West

This paper introduces sketch-guided constrained decoding (SGCD), a novel approach to constrained decoding for blackbox LLMs, which operates without access to the logits of the blackbox LLM.

Constituency Parsing Language Modeling +2

Learning DAGs from Data with Few Root Causes

1 code implementation NeurIPS 2023 Panagiotis Misiakos, Chris Wendler, Markus Püschel

We prove identifiability in this new setting and show that the true DAG is the global minimizer of the $L^0$-norm of the vector of root causes.

Causal Fourier Analysis on Directed Acyclic Graphs and Posets

no code implementations16 Sep 2022 Bastian Seifert, Chris Wendler, Markus Püschel

Specifically, we model the spread of an infection on such a DAG obtained from real-world contact tracing data and learn the infection signal from samples assuming sparsity in the Fourier domain.

Instance-wise algorithm configuration with graph neural networks

1 code implementation10 Feb 2022 Romeo Valentin, Claudio Ferrari, Jérémy Scheurer, Andisheh Amrollahi, Chris Wendler, Max B. Paulus

We pose this task as a supervised learning problem: First, we compile a large dataset of the solver performance for various configurations and all provided MILP instances.

Combinatorial Optimization Graph Neural Network

Learning Set Functions that are Sparse in Non-Orthogonal Fourier Bases

3 code implementations1 Oct 2020 Chris Wendler, Andisheh Amrollahi, Bastian Seifert, Andreas Krause, Markus Püschel

Many applications of machine learning on discrete domains, such as learning preference functions in recommender systems or auctions, can be reduced to estimating a set function that is sparse in the Fourier domain.

Recommendation Systems

Discrete Signal Processing with Set Functions

no code implementations28 Jan 2020 Markus Püschel, Chris Wendler

Set functions are functions (or signals) indexed by the powerset (set of all subsets) of a finite set N. They are fundamental and ubiquitous in many application domains and have been used, for example, to formally describe or quantify loss functions for semantic image segmentation, the informativeness of sensors in sensor networks the utility of sets of items in recommender systems, cooperative games in game theory, or bidders in combinatorial auctions.

Image Segmentation Informativeness +2

Powerset Convolutional Neural Networks

1 code implementation NeurIPS 2019 Chris Wendler, Dan Alistarh, Markus Püschel

We present a novel class of convolutional neural networks (CNNs) for set functions, i. e., data indexed with the powerset of a finite set.

Cannot find the paper you are looking for? You can Submit a new open access paper.