no code implementations • 30 Nov 2024 • Yu Shi, Yiqi Wang, WeiXuan Lang, Jiaxin Zhang, Pan Dong, Aiping Li
Nodes in the real-world graphs exhibit diverse patterns in numerous aspects, such as degree and homophily.
1 code implementation • 20 Oct 2024 • Weichao Zhou, Jiaxin Zhang, Hilaf Hasson, Anu Singh, Wenchao Li
A standard approach is to measure this relevance through the similarity between contexts and queries in the embedding space.
no code implementations • 12 Oct 2024 • Ankita Sinha, Wendi Cui, Kamalika Das, Jiaxin Zhang
Large language models (LLMs) have demonstrated remarkable capabilities; however, the optimization of their prompts has historically prioritized performance metrics at the expense of crucial safety and security considerations.
1 code implementation • 12 Oct 2024 • Jiaxin Zhang, Wendi Cui, Yiran Huang, Kamalika Das, Sricharan Kumar
Large language models (LLMs) are proficient in capturing factual knowledge across various domains.
no code implementations • 10 Oct 2024 • Zhuohang Li, Jiaxin Zhang, Chao Yan, Kamalika Das, Sricharan Kumar, Murat Kantarcioglu, Bradley A. Malin
Retrieval augmented generation (RAG) that retrieves verifiable information from an external knowledge corpus to complement the parametric knowledge in LMs provides a tangible solution to these problems.
no code implementations • 10 Oct 2024 • Xiao Cai, Pengpeng Zeng, Lianli Gao, Junchen Zhu, Jiaxin Zhang, Sitong Su, Heng Tao Shen, Jingkuan Song
Recent advancements in generic 3D content generation from text prompts have been remarkable by fine-tuning text-to-image diffusion (T2I) models or employing these T2I models as priors to learn a general text-to-3D model.
no code implementations • 4 Aug 2024 • Yujin Ren, Jiaxin Zhang, Lianwen Jin
To tackle this problem, a highly promising approach is to utilize massive amounts of unlabeled real data for self-supervised training, which has been widely proven effective in many NLP and CV tasks.
1 code implementation • 31 Jul 2024 • Shiyuan Chen, Jiaxin Zhang, Ruohong Mei, Yingfeng Cai, Haoran Yin, Tao Chen, Wei Sui, Cong Yang
Compared with the original nuScenes static map element, our CAMAv2 annotations achieve lower reprojection errors (e. g., 4. 96 vs. 8. 03 pixels).
no code implementations • 29 Jul 2024 • Jiaxin Zhang, Yunqin Li, Tomohiro Fukuda, Bowen Wang
However, achieving this goal often requires extensive human annotation to train safety ranking models, and the architectural differences between cities hinder the transferability of these models.
no code implementations • 4 Jul 2024 • Zhiyang Xu, Minqian Liu, Ying Shen, Joy Rimchala, Jiaxin Zhang, Qifan Wang, Yu Cheng, Lifu Huang
Lateralization LoRA employs a hybrid approach, combining the traditional linear LoRA and a Convolutional LoRA for generating text and images, enabling the generation of high-quality text and images by leveraging modality-specific structures and parameter sets.
no code implementations • 27 Jun 2024 • Jiaxin Zhang, Wentao Yang, Songxuan Lai, Zecheng Xie, Lianwen Jin
Current multimodal large language models (MLLMs) face significant challenges in visual document understanding (VDU) tasks due to the high resolution, dense text, and complex layouts typical of document images.
no code implementations • 20 Jun 2024 • Minqian Liu, Zhiyang Xu, Zihao Lin, Trevor Ashby, Joy Rimchala, Jiaxin Zhang, Lifu Huang
InterleavedBench features a rich array of tasks to cover diverse real-world use cases.
1 code implementation • CVPR 2024 • Jiaxin Zhang, Dezhi Peng, Chongyu Liu, Peirong Zhang, Lianwen Jin
This underscores the potential of DocRes across a broader spectrum of document image restoration tasks.
1 code implementation • 1 May 2024 • Jiaxin Zhang, Yashar Moshfeghi
Addressing the challenge of automated geometry math problem-solving in artificial intelligence (AI) involves understanding multi-modal information and mathematics.
Ranked #1 on Mathematical Reasoning on UniGeo
no code implementations • 21 Apr 2024 • Jiaxin Zhang, Yiqi Wang, Xihong Yang, Siwei Wang, Yu Feng, Yu Shi, Ruicaho Ren, En Zhu, Xinwang Liu
Graph Neural Networks have demonstrated great success in various fields of multimedia.
1 code implementation • 22 Mar 2024 • Chenyao Yu, Yingfeng Cai, Jiaxin Zhang, Hui Kong, Wei Sui, Cong Yang
As a part of the perception results of intelligent driving systems, static object detection (SOD) in 3D space provides crucial cues for driving environment understanding.
no code implementations • 4 Mar 2024 • Xiang Gao, Jiaxin Zhang, Lalla Mouatadid, Kamalika Das
Motivated by this gap, we introduce a novel UQ method, sampling with perturbation for UQ (SPUQ), designed to tackle both aleatoric and epistemic uncertainties.
no code implementations • 20 Feb 2024 • Jiaxin Zhang, Kamalika Das, Sricharan Kumar
Uncertainty estimation is a crucial aspect of deploying dependable deep learning models in safety-critical systems.
no code implementations • 17 Feb 2024 • Wendi Cui, Jiaxin Zhang, Zhuohang Li, Hao Sun, Damien Lopez, Kamalika Das, Bradley Malin, Sricharan Kumar
Crafting an ideal prompt for Large Language Models (LLMs) is a challenging task that demands significant resources and expert human input.
1 code implementation • 15 Feb 2024 • Jiaxin Zhang, Zhongzhi Li, Mingliang Zhang, Fei Yin, ChengLin Liu, Yashar Moshfeghi
To address this gap, we introduce the GeoEval benchmark, a comprehensive collection that includes a main subset of 2, 000 problems, a 750 problems subset focusing on backward reasoning, an augmented subset of 2, 000 problems, and a hard subset of 300 problems.
no code implementations • 29 Jan 2024 • Jiaxin Zhang, Yinghui Jiang, Yashar Moshfeghi
By leveraging these improvements, GAPS showcases remarkable performance in resolving geometry math problems.
Ranked #1 on Mathematical Reasoning on UniGeo (PRV)
no code implementations • 7 Jan 2024 • Syed Irfan Ali Meerza, Luyang Liu, Jiaxin Zhang, Jian Liu
Specifically, we utilize constrained optimization to enforce local fairness on the client side and adopt a fairness-aware clustering-based aggregation on the server to further ensure the global model fairness across different sensitive groups while maintaining high utility.
1 code implementation • 4 Jan 2024 • Wendi Cui, Jiaxin Zhang, Zhuohang Li, Lopez Damien, Kamalika Das, Bradley Malin, Sricharan Kumar
Evaluating the quality and variability of text generated by Large Language Models (LLMs) poses a significant, yet unresolved research challenge.
no code implementations • 5 Dec 2023 • Dezhi Peng, Zhenhua Yang, Jiaxin Zhang, Chongyu Liu, Yongxin Shi, Kai Ding, Fengjun Guo, Lianwen Jin
Without bells and whistles, the experimental results showcase that the proposed method can simultaneously achieve state-of-the-art performance on three tasks with a unified single model, which provides valuable strategies and insights for future research on generalist OCR models.
no code implementations • 20 Nov 2023 • Zheyuan Kuang, Jiaxin Zhang, Yiying Huang, Yunqin Li
Urban renewal and transformation processes necessitate the preservation of the historical urban fabric, particularly in districts known for their architectural and historical significance.
no code implementations • 16 Nov 2023 • Jiaxin Zhang, Joy Rimchala, Lalla Mouatadid, Kamalika Das, Sricharan Kumar
The performance of optical character recognition (OCR) heavily relies on document image quality, which is crucial for automatic document processing and document intelligence.
no code implementations • 16 Nov 2023 • Sirui Bi, Victor Fung, Jiaxin Zhang
This, in turn, facilitates a probabilistic interpretation of observational data for decision-making.
1 code implementation • 3 Nov 2023 • Jiaxin Zhang, Zhuohang Li, Kamalika Das, Bradley A. Malin, Sricharan Kumar
Hallucination detection is a critical step toward understanding the trustworthiness of modern language models (LMs).
1 code implementation • 21 Sep 2023 • Jiaxin Zhang, Shiyuan Chen, Haoran Yin, Ruohong Mei, Xuan Liu, Cong Yang, Qian Zhang, Wei Sui
The recent development of online static map element (a. k. a.
1 code implementation • 12 Sep 2023 • Jingcan Duan, Pei Zhang, Siwei Wang, Jingtao Hu, Hu Jin, Jiaxin Zhang, Haifang Zhou, Xinwang Liu
Finally, the model is refined with the only input of reliable normal nodes and learns a more accurate estimate of normality so that anomalous nodes can be more easily distinguished.
no code implementations • 31 Aug 2023 • Daoguang Zan, Ailun Yu, Bo Shen, Jiaxin Zhang, Taihong Chen, Bing Geng, Bei Chen, Jichuan Ji, Yafen Yao, Yongji Wang, Qianxiang Wang
Results demonstrate that programming languages can significantly improve each other.
3 code implementations • ICCV 2023 • Mingxin Huang, Jiaxin Zhang, Dezhi Peng, Hao Lu, Can Huang, Yuliang Liu, Xiang Bai, Lianwen Jin
To this end, we introduce a new model named Explicit Synergy-based Text Spotting Transformer framework (ESTextSpotter), which achieves explicit synergy by modeling discriminative and interactive features for text detection and recognition within a single decoder.
no code implementations • 27 Jul 2023 • Bo Shen, Jiaxin Zhang, Taihong Chen, Daoguang Zan, Bing Geng, An Fu, Muhan Zeng, Ailun Yu, Jichuan Ji, Jingyang Zhao, Yuenan Guo, Qianxiang Wang
In this paper, we propose a novel RRTF (Rank Responses to align Test&Teacher Feedback) framework, which can effectively and efficiently boost pre-trained large language models for code generation.
1 code implementation • 20 Jun 2023 • Ruohong Mei, Wei Sui, Jiaxin Zhang, Xue Qin, Gang Wang, Tao Peng, Cong Yang
This paper introduces RoMe, a novel framework designed for the robust reconstruction of large-scale road surfaces.
no code implementations • 9 Jun 2023 • Jiaxin Zhang, Bangdong Chen, Hiuyi Cheng, Fengjun Guo, Kai Ding, Lianwen Jin
Furthermore, considering the importance of fine-grained elements in document images, we present a details recurrent refinement module to enhance the output in a high-resolution space.
no code implementations • 15 May 2023 • Hiuyi Cheng, Peirong Zhang, Sihang Wu, Jiaxin Zhang, Qiyuan Zhu, Zecheng Xie, Jing Li, Kai Ding, Lianwen Jin
Document layout analysis is a crucial prerequisite for document understanding, including document retrieval and conversion.
no code implementations • 11 Apr 2023 • Yue Cui, Syed Irfan Ali Meerza, Zhuohang Li, Luyang Liu, Jiaxin Zhang, Jian Liu
In this paper, we seek to reconcile utility and privacy in FL by proposing a user-configurable privacy defense, RecUP-FL, that can better focus on the user-specified sensitive attributes while obtaining significant improvements in utility over traditional defenses.
no code implementations • 21 Feb 2023 • Zhuohang Li, Jiaxin Zhang, Jian Liu
Distributed machine learning paradigms, such as federated learning, have been recently adopted in many privacy-critical applications for speech analysis.
3 code implementations • 4 Jan 2023 • Yuliang Liu, Jiaxin Zhang, Dezhi Peng, Mingxin Huang, Xinyu Wang, Jingqun Tang, Can Huang, Dahua Lin, Chunhua Shen, Xiang Bai, Lianwen Jin
Within the context of our SPTS v2 framework, our experiments suggest a potential preference for single-point representation in scene text spotting when compared to other representations.
Ranked #15 on Text Spotting on ICDAR 2015
1 code implementation • CVPR 2023 • Hiuyi Cheng, Peirong Zhang, Sihang Wu, Jiaxin Zhang, Qiyuan Zhu, Zecheng Xie, Jing Li, Kai Ding, Lianwen Jin
Document layout analysis is a crucial prerequisite for document understanding, including document retrieval and conversion.
1 code implementation • 8 Dec 2022 • Jiaxin Zhang, Wei Sui, Qian Zhang, Tao Chen, Cong Yang
In this paper, we introduce a novel approach for ground plane normal estimation of wheeled vehicles.
1 code implementation • 2 Dec 2022 • Jiaxin Zhang, Sirui Bi, Victor Fung
In the scope of "AI for Science", solving inverse problems is a longstanding challenge in materials and drug discovery, where the goal is to determine the hidden structures given a set of desirable properties.
1 code implementation • 18 Oct 2022 • Jiaxin Zhang, Yashar Moshfeghi
Numerical reasoning over text is a challenging task of Artificial Intelligence (AI), requiring reading comprehension and numerical reasoning abilities.
Ranked #1 on Math Word Problem Solving on MathQA
1 code implementation • 27 Jul 2022 • Victor Fung, Shuyi Jia, Jiaxin Zhang, Sirui Bi, Junqi Yin, P. Ganesh
These methods would help identify or, in the case of generative models, even create novel crystal structures of materials with a set of specified functional properties to then be synthesized or isolated in the laboratory.
1 code implementation • 23 Jul 2022 • Jiaxin Zhang, Canjie Luo, Lianwen Jin, Fengjun Guo, Kai Ding
To address this issue, we propose a novel approach called Marior (Margin Removal and \Iterative Content Rectification).
2 code implementations • CVPR 2022 • Zhuohang Li, Jiaxin Zhang, Luyang Liu, Jian Liu
Federated Learning (FL) framework brings privacy benefits to distributed learning systems by allowing multiple clients to participate in a learning task under the coordination of a central server without exchanging their private data.
no code implementations • 16 Dec 2021 • Wei Sui, Teng Chen, Jiaxin Zhang, Jiao Lu, Qian Zhang
The Depth-CNN and Pose-CNN estimate dense depth map and ego-motion respectively, solving SFM, while the Pose-CNN and Ground-CNN followed by a homography layer solve the ground plane estimation problem.
1 code implementation • 15 Dec 2021 • Dezhi Peng, Xinyu Wang, Yuliang Liu, Jiaxin Zhang, Mingxin Huang, Songxuan Lai, Shenggao Zhu, Jing Li, Dahua Lin, Chunhua Shen, Xiang Bai, Lianwen Jin
For the first time, we demonstrate that training scene text spotting models can be achieved with an extremely low-cost annotation of a single-point for each instance.
Ranked #3 on Text Spotting on SCUT-CTW1500
no code implementations • 6 Dec 2021 • Alexander Lavin, David Krakauer, Hector Zenil, Justin Gottschlich, Tim Mattson, Johann Brehmer, Anima Anandkumar, Sanjay Choudry, Kamil Rocki, Atılım Güneş Baydin, Carina Prunkl, Brooks Paige, Olexandr Isayev, Erik Peterson, Peter L. McMahon, Jakob Macke, Kyle Cranmer, Jiaxin Zhang, Haruko Wainwright, Adi Hanuka, Manuela Veloso, Samuel Assefa, Stephan Zheng, Avi Pfeffer
We present the "Nine Motifs of Simulation Intelligence", a roadmap for the development and integration of the essential algorithms necessary for a merger of scientific computing, scientific simulation, and artificial intelligence.
no code implementations • NeurIPS 2021 • Ján Drgoňa, Sayak Mukherjee, Jiaxin Zhang, Frank Liu, Mahantesh Halappanavar
Deep Markov models (DMM) are generative models that are scalable and expressive generalization of Markov models for representation, learning, and inference problems.
no code implementations • 29 Sep 2021 • Yu Wang, Jan Drgona, Jiaxin Zhang, Karthik Somayaji NS, Frank Y Liu, Malachi Schram, Peng Li
Although various flow models based on different transformations have been proposed, there still lacks a quantitative analysis of performance-cost trade-offs between different flows as well as a systematic way of constructing the best flow architecture.
no code implementations • 3 Jul 2021 • Zhuohang Li, Luyang Liu, Jiaxin Zhang, Jian Liu
Federated Learning (FL) enables multiple distributed clients (e. g., mobile devices) to collaboratively train a centralized model while keeping the training data locally on the client.
1 code implementation • 6 Jun 2021 • Victor Fung, Jiaxin Zhang, Guoxiang Hu, P. Ganesh, Bobby G. Sumpter
The ability to readily design novel materials with chosen functional properties on-demand represents a next frontier in materials discovery.
no code implementations • 18 Mar 2021 • Jiaxin Zhang, Wei Sui, Xinggang Wang, Wenming Meng, Hongmei Zhu, Qian Zhang
Second, the poses predicted by CNNs are further improved by minimizing photometric errors via gradient updates of poses during inference phases.
no code implementations • 14 Mar 2021 • Jiaxin Zhang, Sirui Bi, Guannan Zhang
However, the approach in Kleinegesse et al., 2020 requires a pathwise sampling path to compute the gradient of the MI lower bound with respect to the design variables, and such a pathwise sampling path is usually inaccessible for implicit models.
no code implementations • 14 Mar 2021 • Jiaxin Zhang, Sirui Bi, Guannan Zhang
However, the approach requires a sampling path to compute the pathwise gradient of the MI lower bound with respect to the design variables, and such a pathwise gradient is usually inaccessible for implicit models.
1 code implementation • 24 Jan 2021 • Jiapeng Wang, Chongyu Liu, Lianwen Jin, Guozhi Tang, Jiaxin Zhang, Shuaitao Zhang, Qianying Wang, Yaqiang Wu, Mingxiang Cai
Visual information extraction (VIE) has attracted considerable attention recently owing to its various advanced applications such as document understanding, automatic marking and intelligent education.
no code implementations • 28 Nov 2020 • Jiaxin Zhang, Congjie Wei, Chenglin Wu
In this paper, we propose a thermodynamic consistent neural network (TCNN) approach to build a data-driven model of the TSR with sparse experimental data.
no code implementations • 28 Nov 2020 • Sirui Bi, Jiaxin Zhang, Guannan Zhang
Unlike the existing studies of DL for TO, our framework accelerates TO by learning the iterative history data and simultaneously training on the mapping between the given design and its gradient.
1 code implementation • 7 Feb 2020 • Jiaxin Zhang, Hoang Tran, Dan Lu, Guannan Zhang
Standard ES methods with $d$-dimensional Gaussian smoothing suffer from the curse of dimensionality due to the high variance of Monte Carlo (MC) based gradient estimators.
no code implementations • 10 Aug 2019 • Jiaxin Zhang, Xianglin Liu, Sirui Bi, Junqi Yin, Guannan Zhang, Markus Eisenbach
In this study, a robust data-driven framework based on Bayesian approaches is proposed and demonstrated on the accurate and efficient prediction of configurational energy of high entropy alloys.
no code implementations • NeurIPS 2019 • Guannan Zhang, Jiaxin Zhang, Jacob Hinkle
We developed a Nonlinear Level-set Learning (NLL) method for dimensionality reduction in high-dimensional function approximation with small data.
Functional Analysis
no code implementations • 26 Feb 2018 • Jinglan Liu, Jiaxin Zhang, Yukun Ding, Xiaowei Xu, Meng Jiang, Yiyu Shi
This work explores the binarization of the deconvolution-based generator in a GAN for memory saving and speedup of image construction.
no code implementations • 29 Oct 2014 • Xiangyang Zhou, Jiaxin Zhang, Brian Kulis
Despite strong performance for a number of clustering tasks, spectral graph cut algorithms still suffer from several limitations: first, they require the number of clusters to be known in advance, but this information is often unknown a priori; second, they tend to produce clusters with uniform sizes.