Search Results for author: Jingxuan Wei

Found 9 papers, 4 papers with code

Sentence-Level or Token-Level? A Comprehensive Study on Knowledge Distillation

no code implementations23 Apr 2024 Jingxuan Wei, Linzhuang Sun, Yichong Leng, Xu Tan, Bihui Yu, Ruifeng Guo

To substantiate our hypothesis, we systematically analyze the performance of distillation methods by varying the model size of student models, the complexity of text, and the difficulty of decoding procedure.

mChartQA: A universal benchmark for multimodal Chart Question Answer based on Vision-Language Alignment and Reasoning

no code implementations2 Apr 2024 Jingxuan Wei, Nan Xu, Guiyong Chang, Yin Luo, Bihui Yu, Ruifeng Guo

In the fields of computer vision and natural language processing, multimodal chart question-answering, especially involving color, structure, and textless charts, poses significant challenges.

Chart Question Answering Language Modelling +1

Unraveling Key Factors of Knowledge Distillation

no code implementations14 Dec 2023 Jingxuan Wei, Linzhuang Sun, Xu Tan, Bihui Yu, Ruifeng Guo

Knowledge distillation, a technique for model compression and performance enhancement, has gained significant traction in Neural Machine Translation (NMT).

Knowledge Distillation Machine Translation +3

Human-computer Interaction for Brain-inspired Computing Based on Machine Learning And Deep Learning: A Review

1 code implementation12 Dec 2023 Bihui Yu, Sibo Zhang, Lili Zhou, Jingxuan Wei, Linzhuang Sun, Liping Bu

Focusing on the application scenarios of decoding text and speech from brain signals in human-computer interaction, this paper presents a comprehensive review of the brain-inspired computing models based on machine learning (ML) and deep learning (DL), tracking their evolution, application value, challenges and potential research trends.

Boosting the Power of Small Multimodal Reasoning Models to Match Larger Models with Self-Consistency Training

1 code implementation23 Nov 2023 Cheng Tan, Jingxuan Wei, Zhangyang Gao, Linzhuang Sun, Siyuan Li, Xihong Yang, Stan Z. Li

Remarkably, we show that even smaller base models, when equipped with our proposed approach, can achieve results comparable to those of larger models, illustrating the potential of our approach in harnessing the power of rationales for improved multimodal reasoning.

Multimodal Reasoning

A Survey on Image-text Multimodal Models

1 code implementation23 Sep 2023 Ruifeng Guo, Jingxuan Wei, Linzhuang Sun, Bihui Yu, Guiyong Chang, Dawei Liu, Sibo Zhang, Zhengbing Yao, Mingjun Xu, Liping Bu

Amidst the evolving landscape of artificial intelligence, the convergence of visual and textual information has surfaced as a crucial frontier, leading to the advent of image-text multimodal models.

Enhancing Human-like Multi-Modal Reasoning: A New Challenging Dataset and Comprehensive Framework

1 code implementation24 Jul 2023 Jingxuan Wei, Cheng Tan, Zhangyang Gao, Linzhuang Sun, Siyuan Li, Bihui Yu, Ruifeng Guo, Stan Z. Li

Multimodal reasoning is a critical component in the pursuit of artificial intelligence systems that exhibit human-like intelligence, especially when tackling complex tasks.

Contrastive Learning Multimodal Reasoning +2

DialogPaint: A Dialog-based Image Editing Model

no code implementations17 Mar 2023 Jingxuan Wei, Shiyu Wu, Xin Jiang, Yequan Wang

We introduce DialogPaint, a novel framework that bridges conversational interactions with image editing, enabling users to modify images through natural dialogue.

Style Transfer

Cannot find the paper you are looking for? You can Submit a new open access paper.