no code implementations • 7 Jan 2024 • Julia El Zini, Bassel Musharrafieh, Mariette Awad
In this work, we investigate the features that are learned by deep models and we study whether these deep networks are able to encode features as complex and high-level as the fractal dimensions.
no code implementations • 5 Feb 2023 • Jihan Ghanim, Maha Issa, Mariette Awad
Building an accurate load forecasting model with minimal underpredictions is vital to prevent any undesired power outages due to underproduction of electricity.
no code implementations • 25 Jan 2023 • Zahraa Al Sahili, Mariette Awad
These powerful algorithms expanded deep learning models to non-Euclidean space and were able to achieve state of art performance in various applications including recommender systems and social networks.
1 code implementation • 19 Jan 2023 • Julia El Zini, Mohammad Mansour, Mariette Awad
Our Contrastive Entropy-based explanation method, CEnt, approximates a model locally by a decision tree to compute entropy information of different feature splits.
1 code implementation • 17 Oct 2022 • Julia El Zini, Mariette Awad
Accordingly, we extend the computation of three metrics, proximity, connectedness and stability, to textual data and we benchmark two successful contrastive methods, POLYJUICE and MiCE, on our suggested metrics.
no code implementations • 13 Oct 2022 • Julia El Zini, Mariette Awad
Such challenges can be attributed to the lack of input structure in textual data, the use of word embeddings that add to the opacity of the models and the difficulty of the visualization of the inner workings of deep models when they are trained on textual data.
no code implementations • 13 Oct 2022 • Julia El Zini, Mohamad Mansour, Basel Mousi, Mariette Awad
In this work, inspired by offline information retrieval, we propose different metrics and techniques to evaluate the explainability of SA models from two angles.
no code implementations • 8 Jul 2022 • Zahraa Al Sahili, Mariette Awad
Additionally, all proposed models were found to accurately classify the 423 classes of plant species, diseases, pests, and weeds with a minimum accuracy of 87% for the Inception-v3 model. Finally, experiments to evaluate of superiority of AgriNet models compared to ImageNet models were conducted on two external datasets: pest and plant diseases dataset from Bangladesh and a plant diseases dataset from Kashmir.
1 code implementation • 25 Jan 2022 • Ali Ismail, Mariette Awad
In the aftermath of disasters, building damage maps are obtained using change detection to plan rescue operations.
1 code implementation • 25 Jan 2022 • Ali Ismail, Mariette Awad
While convolutional neural networks are at the core of recent change detection solutions, we present in this work, BLDNet, a novel graph formulation for building damage change detection and enable learning relationships and representations from both local patterns and non-stationary neighborhoods.
no code implementations • 2 Mar 2021 • Antoine Marot, Benjamin Donnot, Gabriel Dulac-Arnold, Adrian Kelly, Aïdan O'Sullivan, Jan Viebahn, Mariette Awad, Isabelle Guyon, Patrick Panciatici, Camilo Romero
Motivated to investigate the potential of AI methods in enabling adaptability in power network operation, we have designed a L2RPN challenge to encourage the development of reinforcement learning solutions to key problems present in the next-generation power networks.
no code implementations • 19 Jul 2020 • Ariel Ruiz-Garcia, Vasile Palade, Mark Elshaw, Mariette Awad
In this work, we propose a novel Generative Adversarial Stacked Autoencoder that learns to map facial expressions, with up to plus or minus 60 degrees, to an illumination invariant facial representation of 0 degrees.
no code implementations • 26 Nov 2019 • Julia El Zini, Yara Rizk, Mariette Awad
Recurrent neural networks (RNN) have been successfully applied to various sequential decision-making tasks, natural language processing applications, and time-series predictions.
no code implementations • 7 Nov 2019 • Rema Daher, Mohammad Kassem Zein, Julia El Zini, Mariette Awad, Daniel Asmar
Transfer learning improves the GV by 35% and the MS by 13% on average.