no code implementations • 18 Mar 2024 • Angelos Chatzimparmpas, Kostiantyn Kucher, Andreas Kerren
Visualization for explainable and trustworthy machine learning remains one of the most important and heavily researched fields within information visualization and visual analytics with various application domains, such as medicine, finance, and bioinformatics.
1 code implementation • 10 Feb 2024 • Parisa Salmanian, Angelos Chatzimparmpas, Ali Can Karaca, Rafael M. Martins
This paper presents DimVis, a visualization tool that employs supervised Explainable Boosting Machine (EBM) models (trained on user-selected data of interest) as an interpretation assistant for DR projections.
no code implementations • 16 Jan 2024 • Dongping Zhang, Angelos Chatzimparmpas, Negar Kamali, Jessica Hullman
As deep neural networks are more commonly deployed in high-stakes domains, their black-box nature makes uncertainty quantification challenging.
no code implementations • 30 Nov 2023 • Jake M. Hofman, Angelos Chatzimparmpas, Amit Sharma, Duncan J. Watts, Jessica Hullman
Amid rising concerns of reproducibility and generalizability in predictive modeling, we explore the possibility and potential benefits of introducing pre-registration to the field.
1 code implementation • 31 Mar 2023 • Angelos Chatzimparmpas, Rafael M. Martins, Alexandru C. Telea, Andreas Kerren
As the complexity of machine learning (ML) models increases and their application in different (and critical) domains grows, there is a strong demand for more interpretable and trustworthy ML.
no code implementations • 7 Dec 2022 • Ilya Ploshchik, Angelos Chatzimparmpas, Andreas Kerren
Stacking (or stacked generalization) is an ensemble learning method with one main distinctiveness from the rest: even though several base models are trained on the original data set, their predictions are further used as input data for one or more metamodels arranged in at least one extra layer.
1 code implementation • 29 Mar 2022 • Angelos Chatzimparmpas, Fernando V. Paulovich, Andreas Kerren
Our proposed system assists users in visually comparing different distributions of data types, selecting types of instances based on local characteristics that will later be affected by the active sampling method, and validating which suggestions from undersampling or oversampling techniques are beneficial for the ML model.
1 code implementation • 1 Dec 2021 • Angelos Chatzimparmpas, Rafael M. Martins, Andreas Kerren
Bagging and boosting are two popular ensemble methods in machine learning (ML) that produce many individual decision trees.
no code implementations • 26 Mar 2021 • Angelos Chatzimparmpas, Rafael M. Martins, Kostiantyn Kucher, Andreas Kerren
Despite that, while several visual analytics tools exist to monitor and control the different stages of the ML life cycle (especially those related to data and algorithms), feature engineering support remains inadequate.
1 code implementation • 2 Dec 2020 • Angelos Chatzimparmpas, Rafael M. Martins, Kostiantyn Kucher, Andreas Kerren
The challenge is exacerbated by the fact that most ML models are complex internally, and training involves trial-and-error processes that could remarkably affect the predictive result.
1 code implementation • 4 May 2020 • Angelos Chatzimparmpas, Rafael M. Martins, Kostiantyn Kucher, Andreas Kerren
Our system, StackGenVis, assists users in dynamically adapting performance metrics, managing data instances, selecting the most important features for a given data set, choosing a set of top-performant and diverse algorithms, and measuring the predictive performance.
1 code implementation • 17 Feb 2020 • Angelos Chatzimparmpas, Rafael M. Martins, Andreas Kerren
t-Distributed Stochastic Neighbor Embedding (t-SNE) for the visualization of multidimensional data has proven to be a popular approach, with successful applications in a wide range of domains.