1 code implementation • 7 Sep 2024 • Tom Overman, Diego Klabjan, Jean Utke
Automated feature engineering (AutoFE) is the process of automatically building and selecting new features that help improve downstream predictive performance.
no code implementations • 15 Apr 2024 • Jaeyeon Jang, Diego Klabjan, Veena Mendiratta, Fanfei Meng
Federated learning is an emerging paradigm for decentralized training of machine learning models on distributed clients, without revealing the data to the central server.
no code implementations • 6 Feb 2024 • Mengfan Xu, Diego Klabjan
To this end, we are the first to incorporate advanced techniques from blockchains, as well as novel mechanisms, into such a cooperative decision making framework to design optimal strategies for honest participants.
no code implementations • 7 Nov 2023 • Hanqing Li, Diego Klabjan, Jean Utke
This paper introduces a new, unsupervised method for automatic video summarization using ideas from generative adversarial networks but eliminating the discriminator, having a simple loss function, and separating training of different parts of the model.
no code implementations • 5 Nov 2023 • Siqiao Mu, Diego Klabjan
Since the objective functions of reinforcement learning problems are typically highly nonconvex, it is desirable that policy gradient, the most popular algorithm, escapes saddle points and arrives at second-order stationary points.
no code implementations • 3 Nov 2023 • Garrett Blum, Ryan Doris, Diego Klabjan, Horacio Espinosa, Ron Szalkowski
Stress-strain curves, or more generally, stress functions, are an extremely important characterization of a material's mechanical properties.
no code implementations • 16 Oct 2023 • Taejong Joo, Diego Klabjan
Distribution shifts pose significant challenges for model calibration and model selection tasks in the unsupervised domain adaptation problem -- a scenario where the goal is to perform well in a distribution shifted domain without labels.
no code implementations • 15 Aug 2023 • Mengfan Xu, Diego Klabjan
Multi-armed Bandit motivates methods with provable upper bounds on regret and also the counterpart lower bounds have been extensively studied in this context.
no code implementations • 27 Jul 2023 • Shuyang Wang, Diego Klabjan
We propose an ensemble method to improve the generalization performance of trading strategies trained by deep reinforcement learning algorithms in a highly stochastic environment of intraday cryptocurrency portfolio trading.
no code implementations • 13 Jul 2023 • Jaeyeon Jang, Diego Klabjan, Han Liu, Nital S. Patel, Xiuqi Li, Balakrishnan Ananthanarayanan, Husam Dauod, Tzung-Han Juang
This paper proposes a novel multi-agent reinforcement learning (MARL) method to learn multiple coordinated agents under directed acyclic graph (DAG) constraints.
no code implementations • 1 Jul 2023 • Ye Xue, Diego Klabjan, Jean Utke
In this work, we extend and improve Omninet, an architecture that is capable of handling multiple modalities and tasks at a time, by introducing cross-cache attention, integrating patch embeddings for vision inputs, and supporting structured data.
no code implementations • 2 May 2023 • Alexander Cao, Jean Utke, Diego Klabjan
Often pieces of information are received sequentially over time.
no code implementations • 7 Apr 2023 • Alexander Cao, Jean Utke, Diego Klabjan
Sequences are often not received in their entirety at once, but instead, received incrementally over time, element by element.
no code implementations • 28 Feb 2023 • Andrea Treviño Gavito, Diego Klabjan, Jean Utke
Our proposed frameworks allow joint learning on both kinds of data by integrating the paradigms of boosting models and deep neural networks.
no code implementations • 28 Feb 2023 • Andrea Treviño Gavito, Diego Klabjan, Jean Utke
We propose a graph-oriented attention-based explainability method for tabular data.
no code implementations • 21 Dec 2022 • Sungsoo Lim, Diego Klabjan, Mark Shapiro
Feature acquisition algorithms address the problem of acquiring informative features while balancing the costs of acquisition to improve the learning performances of ML models.
no code implementations • 1 Dec 2022 • Mengfan Xu, Diego Klabjan
We study Pareto optimality in multi-objective multi-armed bandit by providing a formulation of adversarial multi-objective multi-armed bandit and defining its Pareto regrets that can be applied to both stochastic and adversarial settings.
no code implementations • 14 Oct 2022 • Tom Overman, Garrett Blum, Diego Klabjan
Furthermore, we provide experimental results that demonstrate the performance improvements of the algorithm over a commonly used method in federated learning, FedAvg, and an existing hybrid FL algorithm, HyFEM.
no code implementations • 11 Oct 2022 • Ruiqi Wang, Diego Klabjan
Under a variance reduction assumption, we show that an ADAM-type algorithm converges, which means that it is the variance of gradients that causes the divergence of original ADAM.
no code implementations • 1 May 2022 • Ning Wang, Han Liu, Diego Klabjan
We develop an abstractive summarization framework independent of labeled data for multiple heterogeneous documents.
no code implementations • 1 Mar 2022 • Biyi Fang, Jean Utke, Diego Klabjan
Convolutional neural networks (CNNs) and transformers, which are composed of multiple processing layers and blocks to learn the representations of data with multiple abstract levels, are the most successful machine learning models in recent years.
no code implementations • 1 Mar 2022 • Biyi Fang, Kripa Rajshekhar, Diego Klabjan
In the context of text, the neural network provides an overview distribution about side data for the corresponding text, which is the prior distribution in LDA to help perform topic grouping.
no code implementations • 9 Jan 2022 • Alexander Cao, Diego Klabjan, Yuan Luo
Open-set recognition generalizes a classification task by classifying test samples as one of the known classes from training or "unknown."
no code implementations • 16 Nov 2021 • Abdolghani Ebrahimi, Diego Klabjan
The proposed algorithm directs the parameters of the compressed model toward a flatter solution by exploring the spectral radius of Hessian which results in better generalization on unseen data.
1 code implementation • 17 Aug 2021 • Ye Xue, Diego Klabjan, Yuan Luo
Federated learning is a distributed machine learning paradigm where multiple data owners (clients) collaboratively train one machine learning model while keeping data on their own devices.
no code implementations • 6 Jul 2021 • Huiyu Wu, Diego Klabjan
We introduce a new, reliable, and agnostic uncertainty measure for classification tasks called logit uncertainty.
no code implementations • 20 May 2021 • Xin Qian, Diego Klabjan
Neural network pruning techniques reduce the number of parameters without compromising predicting ability of a network.
1 code implementation • 22 Feb 2021 • Adam Sandler, Diego Klabjan, Yuan Luo
We develop a regularization method which finds flat minima during the training of deep neural networks and other machine learning models.
no code implementations • 31 Jan 2021 • Stephanie Ger, Diego Klabjan, Jean Utke
Many models such as Long Short Term Memory (LSTMs), Gated Recurrent Units (GRUs) and transformers have been developed to classify time series data with the assumption that events in a sequence are ordered.
no code implementations • 1 Jan 2021 • Xin Qian, Diego Klabjan
We study mini-batch stochastic gradient descent (SGD) dynamics under linear regression and deep linear networks by focusing on the variance of the gradients only given the initial weights and mini-batch size, which is the first study of this nature.
no code implementations • 22 Dec 2020 • Yiming Xu, Diego Klabjan
In this paper, we tackle the open set domain adaptation problem under the assumption that the source and the target label spaces only partially overlap, and the task becomes when the unknown classes exist, how to detect the target unknown classes and avoid aligning them with the source domain.
no code implementations • 8 Dec 2020 • Yiming Xu, Diego Klabjan
Extensive experiments on structured and unstructured data for different type of data changes establish that our method consistently outperforms the state-of-the-art methods by a large margin.
no code implementations • 29 Sep 2020 • Jaehoon Koo, Diego Klabjan, Jean Utke
In this study, we propose a new framework to solve inverse classification that maximizes the number of perturbed samples subject to a per-feature-budget limits and favorable classification classes of the perturbed samples.
no code implementations • 20 Sep 2020 • Mengfan Xu, Diego Klabjan
In EXP4-RL, we extend EXP4. P from bandit scenarios to reinforcement learning to incentivize exploration by multiple agents, including one high-performing agent, for both efficiency and excellence.
no code implementations • 7 Jun 2020 • Qiang Gao, Zhipeng Luo, Diego Klabjan
To reach these goals, we propose a novel approach named as Continual Learning with Efficient Architecture Search, or CLEAS in short.
no code implementations • 3 Jun 2020 • Alexander Cao, Yuan Luo, Diego Klabjan
In inference, open-set classification is to either classify a sample into a known class from training or reject it as an unknown class.
no code implementations • 29 Apr 2020 • Diego Klabjan, Xiaofeng Zhu
We address two challenges of life-long retraining: catastrophic forgetting and efficient retraining.
no code implementations • 27 Apr 2020 • Xin Qian, Diego Klabjan
The mini-batch stochastic gradient descent (SGD) algorithm is widely used in training machine learning models, in particular deep learning models.
no code implementations • 22 Jan 2020 • Xingyu Wang, Lida Zhang, Diego Klabjan
The challenge is that these need to be specified ahead of knowing the forthcoming documents and the underlying topics.
1 code implementation • 7 Jan 2020 • Xiaofeng Zhu, Diego Klabjan
We encode all of the documents already selected by an RNN model.
1 code implementation • 27 Nov 2019 • Adam Sandler, Diego Klabjan, Yuan Luo
We analyze large, multi-dimensional, sparse counting data sets, finding unsupervised groups to provide unique insights into genetic data.
no code implementations • 25 Sep 2019 • Lida Zhang, Diego Klabjan
Deep recurrent neural networks perform well on sequence data and are the model of choice.
1 code implementation • 12 Aug 2019 • Ye Xue, Diego Klabjan, Yuan Luo
The problem of missing values in multivariable time series is a key challenge in many applications such as clinical data mining.
no code implementations • 6 Jun 2019 • Xiaoyi Liu, Diego Klabjan, Patrick NBless
Automatic data extraction from charts is challenging for two reasons: there exist many relations among objects in a chart, which is not a common consideration in general computer vision problems; and different types of charts may not be processed by the same model.
no code implementations • 25 May 2019 • Xin Qian, Matthew Kennedy, Diego Klabjan
In a recurrent setting, conventional approaches to neural architecture search find and fix a general model for all data samples and time steps.
no code implementations • 23 May 2019 • Cheolmin Kim, Youngseok Kim, Diego Klabjan
In this work, we introduce a new class of optimization problems called scale invariant problems and prove that they can be efficiently solved by scale invariant power iteration (SCI-PI) with a generalized convergence guarantee of power iteration.
no code implementations • 22 May 2019 • Biyi Fang, Diego Klabjan
We cover the convex setting showing the regret of the order of the square root of the size of the window in the constant and dynamic learning rate scenarios.
no code implementations • 7 Mar 2019 • Yiming Xu, Dnyanesh Rajpathak, Ian Gibbs, Diego Klabjan
Ontology learning is non-trivial due to several reasons with limited amount of prior research work that automatically learns a domain specific ontology from data.
1 code implementation • 8 Jan 2019 • Stephanie Ger, Yegna Subramanian Jambunath, Diego Klabjan
In addition to the medical device dataset, we also evaluate the GAN-AE performance on two additional datasets and demonstrate the application of GAN-AE to a sequence-to-sequence task where both synthetic sequence inputs and sequence outputs must be generated.
no code implementations • 6 Dec 2018 • Lida Zhang, Abdolghani Ebrahimi, Diego Klabjan
Deep recurrent neural networks perform well on sequence data and are the model of choice.
no code implementations • 27 Sep 2018 • Alexander Stec, Diego Klabjan, Jean Utke
We also include two types of static (whole sequence level) features, one related to time and one not, which are combined with the encoder output.
no code implementations • 25 Sep 2018 • Jaehoon Koo, Diego Klabjan, Jean Utke
Deep learning models based on CNNs are predominantly used in image classification tasks.
no code implementations • 24 Sep 2018 • Alexander Stec, Diego Klabjan, Jean Utke
We also include two types of static (whole sequence level) features, one related to time and one not, which are combined with the encoder output.
no code implementations • 30 Aug 2018 • Alexander Stec, Diego Klabjan, Jean Utke
We also introduce a method to replace instances that are missing which successfully creates neutral input instances and consistently outperforms standard fill-in methods in real world use cases.
no code implementations • 2 Jul 2018 • Mark Harmon, Diego Klabjan
Recurrent neural networks and sequence to sequence models require a predetermined length for prediction output length.
no code implementations • 5 Jun 2018 • Alexander Stec, Diego Klabjan
With our best model we are able to predict the correct bin for overall crime count with 75. 6% and 65. 3% accuracy for Chicago and Portland, respectively.
no code implementations • 4 May 2018 • Jie Yang, Diego Klabjan
In this paper, we propose an active learning algorithm and models which can gradually learn individual's preference through pairwise comparisons.
no code implementations • 27 Apr 2018 • Yiming Xu, Diego Klabjan
In this paper, we propose two families of models built on a sequence to sequence model and a memory network model to mimic the k-Nearest Neighbors model, which generate a sequence of labels, a sequence of out-of-sample feature vectors and a final label for classification, and thus they could also function as oversamplers.
no code implementations • 25 Apr 2018 • Jaehoon Koo, Diego Klabjan
For better classification generative models are used to initialize the model and model features before training a classifier.
no code implementations • 29 Mar 2018 • Biyi Fang, Diego Klabjan
As the size of modern data sets exceeds the disk and memory capacities of a single computer, machine learning practitioners have resorted to parallel and distributed computing.
no code implementations • 15 Feb 2018 • Papis Wongchaisuwat, Diego Klabjan
The proposed system provides supporting evidence when the statement is tagged as false.
1 code implementation • ICML 2018 • Xingyu Wang, Diego Klabjan
Compared to previous works that decouple agents in the game by assuming optimality in expert strategies, we introduce a new objective function that directly pits experts against Nash Equilibrium strategies, and we design an algorithm to solve for the reward function in the context of inverse reinforcement learning with deep neural networks as model approximations.
1 code implementation • 25 Dec 2017 • Baiyang Wang, Diego Klabjan
Generative adversarial nets (GANs) have been successfully applied to the artificial generation of image data.
no code implementations • 27 Nov 2017 • Matthew Dixon, Diego Klabjan, Lan Wei
The OSTSC package is a powerful oversampling approach for classifying univariant, but multinomial time series data in R. This article provides a brief overview of the oversampling methodology implemented by the package.
no code implementations • 28 Sep 2017 • Cheolmin Kim, Diego Klabjan
We present an algorithm for L1-norm kernel PCA and provide a convergence analysis for it.
1 code implementation • IJCNLP 2017 • Xiaofeng Zhu, Diego Klabjan, Patrick Bless
In this paper, we model the document revision detection problem as a minimum cost branching problem that relies on computing document distances.
1 code implementation • 29 Aug 2017 • Xiaofeng Zhu, Diego Klabjan, Patrick Bless
In this paper, we present hierarchical relationbased latent Dirichlet allocation (hrLDA), a data-driven hierarchical topic model for extracting terminological ontologies from a large number of heterogeneous documents.
no code implementations • 6 Jun 2017 • Yaxiong Zeng, Diego Klabjan
In this work, we design a machine learning based method, online adaptive primal support vector regression (SVR), to model the implied volatility surface (IVS).
no code implementations • NeurIPS 2017 • Chao Qin, Diego Klabjan, Daniel Russo
To overcome this shortcoming, we introduce a simple modification of the expected improvement algorithm.
no code implementations • ICLR 2019 • Yintai Ma, Diego Klabjan
Our proposed DBN algorithm remains the overall structure of the original BN algorithm while introduces a weighted averaging update to some trainable parameters.
no code implementations • 24 Feb 2017 • Mark Harmon, Diego Klabjan
We propose a new methodology of designing activation functions within a neural network at each layer.
no code implementations • 20 Feb 2017 • Baiyang Wang, Diego Klabjan
In information retrieval, learning to rank constructs a machine-based ranking model which given a query, sorts the search results by their degree of relevance or importance to the query.
no code implementations • 16 Feb 2017 • Jie Yang, Sergey Shebalov, Diego Klabjan
Two classic semi-supervised learning algorithms, the expectation maximization algorithm and the cluster-and-label algorithm, have been adapted to our choice modeling problem setting.
no code implementations • 27 Jan 2017 • Young Woong Park, Diego Klabjan
For high dimensional cases, an iterative heuristic algorithm is proposed based on the mathematical programming models and a core set concept, and a randomized version of the algorithm is derived to guarantee convergence to the global optimum.
no code implementations • 20 Jan 2017 • Young Woong Park, Diego Klabjan
We propose a mixed integer programming (MIP) model and iterative algorithms based on topological orders to solve optimization problems with acyclic constraints on a directed graph.
1 code implementation • 31 Oct 2016 • Alexandros Nathan, Diego Klabjan
As the size of modern data sets exceeds the disk and memory capacities of a single computer, machine learning practitioners have resorted to parallel and distributed computing.
no code implementations • 15 Sep 2016 • Mark Harmon, Abdolghani Ebrahimi, Patrick Lucey, Diego Klabjan
In this paper, we predict the likelihood of a player making a shot in basketball from multiagent trajectories.
no code implementations • 10 Sep 2016 • Young Woong Park, Diego Klabjan
Principal component analysis (PCA) is often used to reduce the dimension of data by selecting a few orthonormal vectors that explain most of the variance structure of the data.
no code implementations • 15 Aug 2016 • Baiyang Wang, Diego Klabjan
Unsupervised neural networks, such as restricted Boltzmann machines (RBMs) and deep belief networks (DBNs), are powerful tools for feature selection and pattern recognition tasks.
no code implementations • 12 Jul 2016 • Anders Drachen, Eric Thurston Lundquist, Yungjen Kung, Pranav Simha Rao, Diego Klabjan, Rafet Sifa, Julian Runge
Predicting and improving player retention is crucial to the success of mobile Free-to-Play games.
no code implementations • 5 Jul 2016 • Young Woong Park, Diego Klabjan
We propose a clustering-based iterative algorithm to solve certain optimization problems in machine learning, where we start the algorithm by aggregating the original data, solving the problem on aggregated data, and then in subsequent steps gradually disaggregate the aggregated data.
no code implementations • 5 Jul 2016 • Young Woong Park, Yan Jiang, Diego Klabjan, Loren Williams
We examine the performance of our algorithms on a stock keeping unit (SKU) clustering problem employed in forecasting halo and cannibalization effects in promotions using real-world retail data from a large supermarket chain.
no code implementations • 4 Jul 2016 • Papis Wongchaisuwat, Diego Klabjan, Siddhartha R. Jonnalagadda
In this study, we developed an algorithm to automatically answer health-related questions based on past questions and answers (QA).
no code implementations • 4 Jul 2016 • Baiyang Wang, Diego Klabjan
We consider the problem of modeling temporal textual data taking endogenous and exogenous processes into account.
no code implementations • 29 Mar 2016 • Matthew Dixon, Diego Klabjan, Jin Hoon Bang
Deep neural networks (DNNs) are powerful types of artificial neural networks (ANNs) that use several hidden layers.
no code implementations • 24 Mar 2016 • Nikhil Byanna, Diego Klabjan
Using a variety of statistics related to an offensive lineman's performance, we develop a framework to objectively analyze the overall performance of an individual offensive lineman and determine specific linemen who are overvalued or undervalued relative to their salary.
no code implementations • 24 Mar 2016 • Anders Drachen, Matthew Yancey, John Maguire, Derrek Chu, Iris Yuhui Wang, Tobias Mahlmann, Matthias Schubert, Diego Klabjan
Results indicate that spatio-temporal behavior of MOBA teams is related to team skill, with professional teams having smaller within-team distances and conducting more zone changes than amateur teams.
no code implementations • 24 Mar 2016 • Alexander Lavin, Diego Klabjan
Investigations have been performed into using clustering methods in data mining time-series data from smart meters.
no code implementations • 24 Mar 2016 • Taeheon Jeong, Diego Klabjan, Justin Starren
We first find the home location of a patient, which is then augmented with other sensor data to identify sleep patterns and select communication patterns.
no code implementations • 24 Mar 2016 • Anders Drachen, Joseph Riley, Shawna Baskin, Diego Klabjan
The in-game economies of massively multi-player online games (MMOGs) are complex systems that have to be carefully designed and managed.
no code implementations • 24 Mar 2016 • Eun Hee Ko, Diego Klabjan
In addition to major findings, we apply LDA (Latent Dirichlet Allocations) to the same data and drew latent topics that represent consumers' positive opinions and negative opinions on social media.
no code implementations • 23 Mar 2016 • Papis Wongchaisuwat, Diego Klabjan, John O. McGinnis
We develop predictive models for estimating the likelihood of litigation for patents and the expected time to litigation based on both textual and non-textual features.