no code implementations • 10 May 2024 • David "davidad" Dalrymple, Joar Skalse, Yoshua Bengio, Stuart Russell, Max Tegmark, Sanjit Seshia, Steve Omohundro, Christian Szegedy, Ben Goldhaber, Nora Ammann, Alessandro Abate, Joe Halpern, Clark Barrett, Ding Zhao, Tan Zhi-Xuan, Jeannette Wing, Joshua Tenenbaum
Ensuring that AI systems reliably and robustly avoid harmful or dangerous behaviours is a crucial challenge, especially for AI systems with a high degree of autonomy and general intelligence, or systems used in safety-critical contexts.
1 code implementation • 26 Mar 2024 • Jin Peng Zhou, Charles Staats, Wenda Li, Christian Szegedy, Kilian Q. Weinberger, Yuhuai Wu
Large language models (LLM), such as Google's Minerva and OpenAI's GPT families, are becoming increasingly capable of solving mathematical quantitative reasoning problems.
no code implementations • 8 Mar 2023 • Maciej Mikuła, Szymon Tworkowski, Szymon Antoniak, Bartosz Piotrowski, Albert Qiaochu Jiang, Jin Peng Zhou, Christian Szegedy, Łukasz Kuciński, Piotr Miłoś, Yuhuai Wu
By combining \method with a language-model-based automated theorem prover, we further improve the state-of-the-art proof success rate from $57. 0\%$ to $71. 0\%$ on the PISA benchmark using $4$x fewer parameters.
no code implementations • 25 May 2022 • Yuhuai Wu, Albert Q. Jiang, Wenda Li, Markus N. Rabe, Charles Staats, Mateja Jamnik, Christian Szegedy
Autoformalization is the process of automatically translating from natural language mathematics to formal specifications and proofs.
4 code implementations • ICLR 2022 • Yuhuai Wu, Markus N. Rabe, DeLesley Hutchins, Christian Szegedy
Language models typically need to be trained or finetuned in order to acquire new knowledge, which involves updating their weights.
3 code implementations • Findings (NAACL) 2022 • Piotr Nawrot, Szymon Tworkowski, Michał Tyrolski, Łukasz Kaiser, Yuhuai Wu, Christian Szegedy, Henryk Michalewski
Transformer models yield impressive results on many NLP and sequence modeling tasks.
Ranked #8 on
Image Generation
on ImageNet 32x32
(bpd metric)
1 code implementation • 15 Jan 2021 • Yuhuai Wu, Markus Rabe, Wenda Li, Jimmy Ba, Roger Grosse, Christian Szegedy
While designing inductive bias in neural architectures has been widely studied, we hypothesize that transformer networks are flexible enough to learn inductive bias from suitable generic tasks.
no code implementations • ICLR 2021 • Markus N. Rabe, Dennis Lee, Kshitij Bansal, Christian Szegedy
We examine whether self-supervised language modeling applied to mathematical formulas enables logical reasoning.
no code implementations • ICLR 2020 • Dennis Lee, Christian Szegedy, Markus N. Rabe, Sarah M. Loos, Kshitij Bansal
We design and conduct a simple experiment to study whether neural networks can perform several steps of approximate reasoning in a fixed dimensional latent space.
no code implementations • 25 May 2019 • Kshitij Bansal, Christian Szegedy, Markus N. Rabe, Sarah M. Loos, Viktor Toman
Our experiments show that the theorem prover trained with this exploration mechanism outperforms provers that are trained only on human proofs.
Ranked #3 on
Automated Theorem Proving
on HOList benchmark
no code implementations • 24 May 2019 • Aditya Paliwal, Sarah Loos, Markus Rabe, Kshitij Bansal, Christian Szegedy
This paper presents the first use of graph neural networks (GNNs) for higher-order proof search and demonstrates that GNNs can improve upon state-of-the-art results in this domain.
Ranked #1 on
Automated Theorem Proving
on HOList benchmark
3 code implementations • 5 Apr 2019 • Kshitij Bansal, Sarah M. Loos, Markus N. Rabe, Christian Szegedy, Stewart Wilcox
We present an environment, benchmark, and deep learning driven automated theorem prover for higher-order logic.
Ranked #2 on
Automated Theorem Proving
on HOList benchmark
no code implementations • ICLR 2019 • Tolgahan Cakaloglu, Christian Szegedy, Xiaowei Xu
Text embedding representing natural language documents in a semantic vector space can be used for document retrieval using nearest neighbor lookup.
1 code implementation • 1 Mar 2017 • Cezary Kaliszyk, François Chollet, Christian Szegedy
We propose various machine learning tasks that can be performed on this dataset, and discuss their significance for theorem proving.
Ranked #3 on
Automated Theorem Proving
on HolStep (Unconditional)
no code implementations • 24 Jan 2017 • Sarah Loos, Geoffrey Irving, Christian Szegedy, Cezary Kaliszyk
Here we suggest deep learning based guidance in the proof search of the theorem prover E. We train and compare several deep neural network models on the traces of existing ATP proofs of Mizar statements and use them to select processed clauses during proof search.
2 code implementations • NeurIPS 2016 • Alex A. Alemi, Francois Chollet, Niklas Een, Geoffrey Irving, Christian Szegedy, Josef Urban
We study the effectiveness of neural sequence models for premise selection in automated theorem proving, one of the main bottlenecks in the formalization of mathematics.
86 code implementations • 23 Feb 2016 • Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, Alex Alemi
Recently, the introduction of residual connections in conjunction with a more traditional architecture has yielded state-of-the-art performance in the 2015 ILSVRC challenge; its performance was similar to the latest generation Inception-v3 network.
Ranked #4 on
Classification
on InDL
no code implementations • 17 Dec 2015 • Qian Yu, Christian Szegedy, Martin C. Stumpe, Liron Yatziv, Vinay Shet, Julian Ibarz, Sacha Arnoud
Precise business store front detection enables accurate geo-location of businesses, and further provides input for business categorization, listing generation, etc.
223 code implementations • 8 Dec 2015 • Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, Alexander C. Berg
Experimental results on the PASCAL VOC, MS COCO, and ILSVRC datasets confirm that SSD has comparable accuracy to methods that utilize an additional object proposal step and is much faster, while providing a unified framework for both training and inference.
Ranked #3 on
Object Detection
on PASCAL VOC 2012
112 code implementations • CVPR 2016 • Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jonathon Shlens, Zbigniew Wojna
Convolutional networks are at the core of most state-of-the-art computer vision solutions for a wide variety of tasks.
Ranked #8 on
Retinal OCT Disease Classification
on OCT2017
no code implementations • ICML 2015 2015 • Sergey Ioffe, Christian Szegedy
Training Deep Neural Networks is complicated by the factthat the distribution of each layer’s inputs changes duringtraining, as the parameters of the previous layers change. This slows down the training by requiring lower learningrates and careful parameter initialization, and makes it no-toriously hard to train models with saturating nonlineari-ties.
72 code implementations • 11 Feb 2015 • Sergey Ioffe, Christian Szegedy
Training Deep Neural Networks is complicated by the fact that the distribution of each layer's inputs changes during training, as the parameters of the previous layers change.
3 code implementations • 20 Dec 2014 • Scott Reed, Honglak Lee, Dragomir Anguelov, Christian Szegedy, Dumitru Erhan, Andrew Rabinovich
On MNIST handwritten digits, we show that our model is robust to label corruption.
60 code implementations • 20 Dec 2014 • Ian J. Goodfellow, Jonathon Shlens, Christian Szegedy
Several machine learning models, including neural networks, consistently misclassify adversarial examples---inputs formed by applying small but intentionally worst-case perturbations to examples from the dataset, such that the perturbed input results in the model outputting an incorrect answer with high confidence.
Ranked #45 on
Image Classification
on MNIST
no code implementations • 3 Dec 2014 • Christian Szegedy, Scott Reed, Dumitru Erhan, Dragomir Anguelov, Sergey Ioffe
Using the multi-scale convolutional MultiBox (MSC-MultiBox) approach, we substantially advance the state-of-the-art on the ILSVRC 2014 detection challenge data set, with $0. 5$ mAP for a single model and $0. 52$ mAP for an ensemble of two models.
81 code implementations • CVPR 2015 • Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, Andrew Rabinovich
We propose a deep convolutional neural network architecture codenamed "Inception", which was responsible for setting the new state of the art for classification and detection in the ImageNet Large-Scale Visual Recognition Challenge 2014 (ILSVRC 2014).
12 code implementations • 21 Dec 2013 • Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, Rob Fergus
Deep neural networks are highly expressive models that have recently achieved state of the art performance on speech and visual recognition tasks.
7 code implementations • CVPR 2014 • Alexander Toshev, Christian Szegedy
We propose a method for human pose estimation based on Deep Neural Networks (DNNs).
6 code implementations • CVPR 2014 • Dumitru Erhan, Christian Szegedy, Alexander Toshev, Dragomir Anguelov
Deep convolutional neural networks have recently achieved state-of-the-art performance on a number of image recognition benchmarks, including the ImageNet Large-Scale Visual Recognition Challenge (ILSVRC-2012).
no code implementations • NeurIPS 2013 • Christian Szegedy, Alexander Toshev, Dumitru Erhan
Deep Neural Networks (DNNs) have recently shown outstanding performance on the task of whole image classification.