1 code implementation • 26 Mar 2024 • Jin Peng Zhou, Charles Staats, Wenda Li, Christian Szegedy, Kilian Q. Weinberger, Yuhuai Wu
Large language models (LLM), such as Google's Minerva and OpenAI's GPT families, are becoming increasingly capable of solving mathematical quantitative reasoning problems.
1 code implementation • 26 Feb 2024 • Jin Peng Zhou, Yuhuai Wu, Qiyang Li, Roger Grosse
With newly extracted theorems, we show that the existing proofs in the MetaMath database can be refactored.
no code implementations • 5 Feb 2024 • Quang-Huy Nguyen, Jin Peng Zhou, Zhenzhen Liu, Khanh-Huyen Bui, Kilian Q. Weinberger, Dung D. Le
RONIN conditions the inpainting process with the predicted ID label, drawing the input object closer to the in-distribution domain.
no code implementations • 24 Oct 2023 • Zhenzhen Liu, Chao Wan, Varsha Kishore, Jin Peng Zhou, Minmin Chen, Kilian Q. Weinberger
The results show that CoBa is effective and efficient in reducing hallucination, and offers great adaptability and flexibility.
no code implementations • 8 Mar 2023 • Maciej Mikuła, Szymon Tworkowski, Szymon Antoniak, Bartosz Piotrowski, Albert Qiaochu Jiang, Jin Peng Zhou, Christian Szegedy, Łukasz Kuciński, Piotr Miłoś, Yuhuai Wu
By combining \method with a language-model-based automated theorem prover, we further improve the state-of-the-art proof success rate from $57. 0\%$ to $71. 0\%$ on the PISA benchmark using $4$x fewer parameters.
1 code implementation • 20 Feb 2023 • Zhenzhen Liu, Jin Peng Zhou, Yufan Wang, Kilian Q. Weinberger
We present a novel approach for this task - Lift, Map, Detect (LMD) - that leverages recent advancement in diffusion models.
no code implementations • 20 Dec 2022 • Roei Schuster, Jin Peng Zhou, Thorsten Eisenhofer, Paul Grubbs, Nicolas Papernot
We analyze the root causes of potentially-increased attack surface in learned systems and develop a framework for identifying vulnerabilities that stem from the use of ML.
3 code implementations • 21 Oct 2022 • Albert Q. Jiang, Sean Welleck, Jin Peng Zhou, Wenda Li, Jiacheng Liu, Mateja Jamnik, Timothée Lacroix, Yuhuai Wu, Guillaume Lample
In this work, we introduce Draft, Sketch, and Prove (DSP), a method that maps informal proofs to formal proof sketches, and uses the sketches to guide an automated prover by directing its search to easier sub-problems.
Ranked #3 on Automated Theorem Proving on miniF2F-valid (Pass@100 metric)
1 code implementation • 25 Feb 2022 • Ruihan Wu, Jin Peng Zhou, Kilian Q. Weinberger, Chuan Guo
Label differential privacy (label-DP) is a popular framework for training private ML models on datasets with public features and sensitive private labels.
no code implementations • 20 Aug 2020 • Baiwu Zhang, Jin Peng Zhou, Ilia Shumailov, Nicolas Papernot
We discuss the ethical implications of our work, identify where our technique can be used, and highlight that a more meaningful legislative framework is required for a more transparent and ethical use of generative modeling.
no code implementations • 3 Aug 2020 • Jin Peng Zhou, Ga Wu, Zheda Mai, Scott Sanner
One-class collaborative filtering (OC-CF) is a common class of recommendation problem where only the positive class is explicitly observed (e. g., purchases, clicks).