Search Results for author: Ehsan Imani

Found 8 papers, 3 papers with code

Investigating the Histogram Loss in Regression

1 code implementation20 Feb 2024 Ehsan Imani, Kai Luedemann, Sam Scholnick-Hughes, Esraa Elelimy, Martha White

It is becoming increasingly common in regression to train neural networks that model the entire distribution even if only the mean is required for prediction.

regression

Label Alignment Regularization for Distribution Shift

no code implementations27 Nov 2022 Ehsan Imani, Guojun Zhang, Runjia Li, Jun Luo, Pascal Poupart, Philip H. S. Torr, Yangchen Pan

Recent work has highlighted the label alignment property (LAP) in supervised learning, where the vector of all labels in the dataset is mostly in the span of the top few singular vectors of the data matrix.

Representation Learning Sentiment Analysis +1

Representation Alignment in Neural Networks

1 code implementation15 Dec 2021 Ehsan Imani, Wei Hu, Martha White

We then highlight why alignment between the top singular vectors and the targets can speed up learning and show in a classic synthetic transfer problem that representation alignment correlates with positive and negative transfer to similar and dissimilar tasks.

Off-Policy Actor-Critic with Emphatic Weightings

1 code implementation16 Nov 2021 Eric Graves, Ehsan Imani, Raksha Kumaraswamy, Martha White

A variety of theoretically-sound policy gradient algorithms exist for the on-policy setting due to the policy gradient theorem, which provides a simplified form for the gradient.

Hallucinating Value: A Pitfall of Dyna-style Planning with Imperfect Environment Models

no code implementations8 Jun 2020 Taher Jafferjee, Ehsan Imani, Erin Talvitie, Martha White, Micheal Bowling

Dyna-style reinforcement learning (RL) agents improve sample efficiency over model-free RL agents by updating the value function with simulated experience generated by an environment model.

Reinforcement Learning (RL)

An implicit function learning approach for parametric modal regression

no code implementations NeurIPS 2020 Yangchen Pan, Ehsan Imani, Martha White, Amir-Massoud Farahmand

We empirically demonstrate on several synthetic problems that our method (i) can learn multi-valued functions and produce the conditional modes, (ii) scales well to high-dimensional inputs, and (iii) can even be more effective for certain uni-modal problems, particularly for high-frequency functions.

regression

An Off-policy Policy Gradient Theorem Using Emphatic Weightings

no code implementations NeurIPS 2018 Ehsan Imani, Eric Graves, Martha White

There have been a host of theoretically sound algorithms proposed for the on-policy setting, due to the existence of the policy gradient theorem which provides a simplified form for the gradient.

Policy Gradient Methods

Improving Regression Performance with Distributional Losses

no code implementations ICML 2018 Ehsan Imani, Martha White

We provide theoretical support for this alternative hypothesis, by characterizing the norm of the gradients of this loss.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.