Search Results for author: Audrey Durand

Found 21 papers, 8 papers with code

Randomized Confidence Bounds for Stochastic Partial Monitoring

no code implementations7 Feb 2024 Maxime Heuillet, Ola Ahmad, Audrey Durand

In this paper, we consider the contextual and non-contextual PM settings with stochastic outcomes.

Association Rules Mining with Auto-Encoders

no code implementations26 Apr 2023 Théophile Berteloot, Richard Khoury, Audrey Durand

Classical association rule mining algorithms have several limitations, especially with regards to their high execution times and number of rules produced.

Interpret Your Care: Predicting the Evolution of Symptoms for Cancer Patients

no code implementations19 Feb 2023 Rupali Bhati, Jennifer Jones, Audrey Durand

The focus of this study is on predicting the pain and tiredness level of a patient post their diagnosis.

Neural Bandits for Data Mining: Searching for Dangerous Polypharmacy

1 code implementation10 Dec 2022 Alexandre Larouche, Audrey Durand, Richard Khoury, Caroline Sirois

Polypharmacy, most often defined as the simultaneous consumption of five or more drugs at once, is a prevalent phenomenon in the older population.

Thompson Sampling

Cambrian Explosion Algorithm for Multi-Objective Association Rules Mining

no code implementations23 Nov 2022 Théophile Berteloot, Richard Khoury, Audrey Durand

Association rule mining is one of the most studied research fields of data mining, with applications ranging from grocery basket problems to highly explainable classification systems.

Algorithms for Adaptive Experiments that Trade-off Statistical Analysis with Reward: Combining Uniform Random Assignment and Reward Maximization

no code implementations15 Dec 2021 Tong Li, Jacob Nogas, Haochen Song, Harsh Kumar, Audrey Durand, Anna Rafferty, Nina Deliu, Sofia S. Villar, Joseph J. Williams

TS-PostDiff takes a Bayesian approach to mixing TS and Uniform Random (UR): the probability a participant is assigned using UR allocation is the posterior probability that the difference between two arms is 'small' (below a certain threshold), allowing for more UR exploration when there is little or no reward to be gained.

Thompson Sampling

GrowSpace: Learning How to Shape Plants

no code implementations15 Oct 2021 Yasmeen Hitti, Ionelia Buzatu, Manuel Del Verme, Mark Lefsrud, Florian Golemo, Audrey Durand

We argue that plant responses to an environmental stimulus are a good example of a real-world problem that can be approached within a reinforcement learning (RL)framework.

Fairness Reinforcement Learning (RL)

Challenges in Statistical Analysis of Data Collected by a Bandit Algorithm: An Empirical Exploration in Applications to Adaptively Randomized Experiments

no code implementations22 Mar 2021 Joseph Jay Williams, Jacob Nogas, Nina Deliu, Hammad Shaikh, Sofia S. Villar, Audrey Durand, Anna Rafferty

We therefore use our case study of the ubiquitous two-arm binary reward setting to empirically investigate the impact of using Thompson Sampling instead of uniform random assignment.

Thompson Sampling

Comparison of pharmacist evaluation of medication orders with predictions of a machine learning model

1 code implementation3 Nov 2020 Sophie-Camille Hogue, Flora Chen, Geneviève Brassard, Denis Lebel, Jean-François Bussières, Audrey Durand, Maxime Thibault

The objective of this work was to assess the clinical performance of an unsupervised machine learning model aimed at identifying unusual medication orders and pharmacological profiles.

BIG-bench Machine Learning

Deep interpretability for GWAS

no code implementations3 Jul 2020 Deepak Sharma, Audrey Durand, Marc-André Legault, Louis-Philippe Lemieux Perreault, Audrey Lemaçon, Marie-Pierre Dubé, Joelle Pineau

Genome-Wide Association Studies are typically conducted using linear models to find genetic variants associated with common diseases.

Old Dog Learns New Tricks: Randomized UCB for Bandit Problems

1 code implementation11 Oct 2019 Sharan Vaswani, Abbas Mehrabian, Audrey Durand, Branislav Kveton

We propose $\tt RandUCB$, a bandit strategy that builds on theoretically derived confidence intervals similar to upper confidence bound (UCB) algorithms, but akin to Thompson sampling (TS), it uses randomization to trade off exploration and exploitation.

Thompson Sampling

Attraction-Repulsion Actor-Critic for Continuous Control Reinforcement Learning

no code implementations17 Sep 2019 Thang Doan, Bogdan Mazoure, Moloud Abdar, Audrey Durand, Joelle Pineau, R. Devon Hjelm

Continuous control tasks in reinforcement learning are important because they provide an important framework for learning in high-dimensional state spaces with deceptive rewards, where the agent can easily become trapped into suboptimal solutions.

Continuous Control reinforcement-learning +1

Leveraging exploration in off-policy algorithms via normalizing flows

1 code implementation16 May 2019 Bogdan Mazoure, Thang Doan, Audrey Durand, R. Devon Hjelm, Joelle Pineau

The ability to discover approximately optimal policies in domains with sparse rewards is crucial to applying reinforcement learning (RL) in many real-world scenarios.

Continuous Control Reinforcement Learning (RL)

On-line Adaptative Curriculum Learning for GANs

3 code implementations31 Jul 2018 Thang Doan, Joao Monteiro, Isabela Albuquerque, Bogdan Mazoure, Audrey Durand, Joelle Pineau, R. Devon Hjelm

We argue that less expressive discriminators are smoother and have a general coarse grained view of the modes map, which enforces the generator to cover a wide portion of the data distribution support.

Multi-Armed Bandits Stochastic Optimization

Learning to Become an Expert: Deep Networks Applied To Super-Resolution Microscopy

no code implementations28 Mar 2018 Louis-Émile Robitaille, Audrey Durand, Marc-André Gardner, Christian Gagné, Paul De Koninck, Flavie Lavoie-Cardinal

More specifically, we are proposing a system based on a deep neural network that can provide a quantitative quality measure of a STED image of neuronal structures given as input.

Super-Resolution

Estimating Quality in Multi-Objective Bandits Optimization

no code implementations4 Jan 2017 Audrey Durand, Christian Gagné

The question is: how good do estimations of these objectives have to be in order for the solution maximizing the preference function to remain unchanged?

Thompson Sampling

Cannot find the paper you are looking for? You can Submit a new open access paper.