no code implementations • 3 Oct 2022 • Krishna Dasaratha, Kevin He
We study learning on social media with an equilibrium model of users interacting with shared news stories.
no code implementations • 10 Aug 2022 • Li Liu, Xiangeng Fang, Di Wang, Weijing Tang, Kevin He
Neural Network (Deep Learning) is a modern model in Artificial Intelligence and it has been exploited in Survival Analysis.
no code implementations • 3 Jan 2022 • Daniel Clark, Drew Fudenberg, Kevin He
Learning models do not in general imply that weakly dominated strategies are irrelevant or justify the related concept of "forward induction," because rational agents may use dominated strategies as experiments to learn how opponents play, and may not have enough data to rule out a strategy that opponents never use.
no code implementations • 29 Dec 2021 • Kevin He, Fedor Sandomirskiy, Omer Tamuz
A private private information structure delivers information about an unknown state while preserving privacy: An agent's signal contains information about the state but remains independent of others' sensitive or private information.
no code implementations • 16 Mar 2021 • Federico Echenique, Kevin He
Uninformed $p$-hackers, who are fully ignorant of the true mechanism and engage in data mining, often fall for baits.
no code implementations • 30 Dec 2020 • Kevin He, Jonathan Libgober
Toward explaining the persistence of biased inferences, we propose a framework to evaluate competing (mis)specifications in strategic settings.
no code implementations • 22 Nov 2019 • Krishna Dasaratha, Kevin He
In a class of networks where agents move in generations and observe the previous generation, we quantify the information loss with an aggregative efficiency index.
no code implementations • 5 Sep 2019 • Krishna Dasaratha, Kevin He
A network determines the observable predecessors, and we compare subjects' accuracy on sparse and dense networks.
no code implementations • 31 Jul 2019 • Jetlir Duraj, Kevin He
Diminishing sensitivity induces a preference over news skewness: gradual bad news, one-shot good news is worse than one-shot resolution, which is in turn worse than gradual good news, one-shot bad news.
no code implementations • 17 May 2018 • Kevin He, Jian Kang, Hyokyoung Grace Hong, Ji Zhu, Yanming Li, Huazhen Lin, Han Xu, Yi Li
Modern bio-technologies have produced a vast amount of high-throughput data with the number of predictors far greater than the sample size.
no code implementations • 21 Mar 2018 • Kevin He
When agents wrongly expect systematic reversals (the "gambler's fallacy"), they understate the likelihood of consecutive below-average draws, converge to over-pessimistic beliefs about the distribution's mean, and stop too early.
no code implementations • 4 Nov 2016 • Yanming Li, Hyokyoung Hong, Jian Kang, Kevin He, Ji Zhu, Yi Li
Although much progress has been made in classification with high-dimensional features \citep{Fan_Fan:2008, JGuo:2010, CaiSun:2014, PRXu:2014}, classification with ultrahigh-dimensional features, wherein the features much outnumber the sample size, defies most existing work.