Search Results for author: Giorgio Giannone

Found 12 papers, 4 papers with code

NITO: Neural Implicit Fields for Resolution-free Topology Optimization

no code implementations7 Feb 2024 Amin Heyrani Nobari, Giorgio Giannone, Lyle Regenwetter, Faez Ahmed

We introduce Neural Implicit Topology Optimization (NITO), a novel approach to accelerate topology optimization problems using deep learning.

Learning from Invalid Data: On Constraint Satisfaction in Generative Models

no code implementations27 Jun 2023 Giorgio Giannone, Lyle Regenwetter, Akash Srivastava, Dan Gutfreund, Faez Ahmed

This is particularly problematic when the generated data must satisfy constraints, for example, to meet product specifications in engineering design or to adhere to the laws of physics in a natural scene.

valid

Diffusing the Optimal Topology: A Generative Optimization Approach

no code implementations17 Mar 2023 Giorgio Giannone, Faez Ahmed

To address these issues, we propose a Generative Optimization method that integrates classic optimization like SIMP as a refining mechanism for the topology generated by a deep generative model.

Unifying Molecular and Textual Representations via Multi-task Language Modelling

1 code implementation29 Jan 2023 Dimitrios Christofidellis, Giorgio Giannone, Jannis Born, Ole Winther, Teodoro Laino, Matteo Manica

Here, we propose the first multi-domain, multi-task language model that can solve a wide range of tasks in both the chemical and natural language domains.

Language Modelling Molecule Captioning +2

Just Mix Once: Worst-group Generalization by Group Interpolation

no code implementations21 Oct 2022 Giorgio Giannone, Serhii Havrylov, Jordan Massiah, Emine Yilmaz, Yunlong Jiao

Advances in deep learning theory have revealed how average generalization relies on superficial patterns in data.

Learning Theory

Few-Shot Diffusion Models

1 code implementation30 May 2022 Giorgio Giannone, Didrik Nielsen, Ole Winther

At test time, the model is able to generate samples from previously unseen classes conditioned on as few as 5 samples from that class.

Denoising Few-Shot Learning

SCHA-VAE: Hierarchical Context Aggregation for Few-Shot Generation

1 code implementation23 Oct 2021 Giorgio Giannone, Ole Winther

In few-shot learning the model is trained on data from many sets from distributions sharing some underlying properties such as sets of characters from different alphabets or objects from different categories.

Few-Shot Learning Out-of-Distribution Generalization

No Representation without Transformation

no code implementations9 Dec 2019 Giorgio Giannone, Saeed Saremi, Jonathan Masci, Christian Osendorfer

To explicitly demonstrate the effect of these higher order objects, we show that the inferred latent transformations reflect interpretable properties in the observation space.

Learning Common Representation from RGB and Depth Images

no code implementations17 Dec 2018 Giorgio Giannone, Boris Chidlovskii

We propose a new deep learning architecture for the tasks of semantic segmentation and depth prediction from RGB-D images.

Depth Estimation Depth Prediction +3

Cannot find the paper you are looking for? You can Submit a new open access paper.