1 code implementation • 20 Jan 2025 • Wannita Takerngsaksiri, Micheal Fu, Chakkrit Tantithamthavorn, Jirat Pasuksmit, Kun Chen, Ming Wu
In this paper, we conduct a survey to explore the practitioners' perspectives on code readability in the age of LLMs and investigate the readability of our LLM-based software development agents framework, HULA, by comparing its generated code with human-written code in real-world scenarios.
2 code implementations • 12 Jan 2025 • Shaw Walters, Sam Gao, Shakker Nerd, Feng Da, Warren Williams, Ting-Chien Meng, Amie Chow, Hunter Han, Frank He, Allen Zhang, Ming Wu, Timothy Shen, Maxwell Hu, Jerry Yan
AI Agent, powered by large language models (LLMs) as its cognitive core, is an intelligent agentic system capable of autonomously controlling and determining the execution paths under user's instructions.
no code implementations • 25 Nov 2024 • Xingshuo Han, Xuanye Zhang, Xiang Lan, Haozhao Wang, Shengmin Xu, Shen Ren, Jason Zeng, Ming Wu, Michael Heinrich, Tianwei Zhang
Federated learning (FL) enables the training of deep learning models on distributed clients to preserve data privacy.
no code implementations • 19 Nov 2024 • Wannita Takerngsaksiri, Jirat Pasuksmit, Patanamon Thongtanunam, Chakkrit Tantithamthavorn, Ruixiong Zhang, Fan Jiang, Jing Li, Evan Cook, Kun Chen, Ming Wu
In this paper, we introduce a Human-in-the-loop LLM-based Agents framework (HULA) for software development that allows software engineers to refine and guide LLMs when generating coding plans and source code for a given task.
2 code implementations • 10 Nov 2024 • Jinbo Hu, Yin Cao, Ming Wu, Fang Kang, Feiran Yang, Wenwu Wang, Mark D. Plumbley, Jun Yang
Recently, deep neural networks trained on large-scale datasets have achieved remarkable success in the sound event classification (SEC) field, prompting an open question of whether these advancements can be extended to develop general-purpose SELD models.
Direction of Arrival Estimation
Sound Event Localization and Detection
1 code implementation • 19 Jun 2024 • Mengqiu Xu, Ming Wu, Kaixin Chen, Yixiang Huang, Mingrui Xu, Yujia Yang, Yiqing Feng, Yiying Guo, Bin Huang, Dongliang Chang, Zhenwei Shi, Chuang Zhang, Zhanyu Ma, Jun Guo
Marine fog poses a significant hazard to global shipping, necessitating effective detection and forecasting to reduce economic losses.
2 code implementations • 25 Apr 2024 • Haotian Yan, Ming Wu, Chuang Zhang
VWA leverages the local window attention (LWA) and disentangles LWA into the query window and context window, allowing the context's scale to vary for the query to learn representations at multiple scales.
no code implementations • 26 Jan 2024 • Yutong Xiong, Xun Zhu, Ming Wu, Weiqing Li, Fanbin Mo, Chuang Zhang, Bin Zhang
Sparse meteorological forecasting is indispensable for fine-grained weather forecasting and deserves extensive attention.
1 code implementation • 27 Dec 2023 • Jinbo Hu, Yin Cao, Ming Wu, Qiuqiang Kong, Feiran Yang, Mark D. Plumbley, Jun Yang
In addition, we introduce environment representations to characterize different acoustic settings, enhancing the adaptability of our attenuation approach to various environments.
no code implementations • 17 Aug 2023 • Jinbo Hu, Yin Cao, Ming Wu, Feiran Yang, Ziying Yu, Wenwu Wang, Mark D. Plumbley, Jun Yang
For learning-based sound event localization and detection (SELD) methods, different acoustic environments in the training and test sets may result in large performance differences in the validation and evaluation stages.
1 code implementation • 8 Jul 2023 • Yi Zhong, Mengqiu Xu, Kongming Liang, Kaixin Chen, Ming Wu
Segmentation of the infected areas of the lung is essential for quantifying the severity of lung disease like pulmonary infections.
1 code implementation • 1 Jun 2023 • Banghua Zhu, Mingyu Ding, Philip Jacobson, Ming Wu, Wei Zhan, Michael Jordan, Jiantao Jiao
Self-training is an important technique for solving semi-supervised learning problems.
no code implementations • 26 Feb 2023 • Shenwei Xie, Wanfeng Zheng, Zhenglin Xian, Junli Yang, Chuang Zhang, Ming Wu
In this paper, we propose a new scheme for multi-task satellite imagery road extraction, Patch-wise Road Keypoints Detection (PaRK-Detect).
1 code implementation • 21 Feb 2023 • Xun Zhu, Yutong Xiong, Ming Wu, Gaozhen Nie, Bin Zhang, Ziheng Yang
To the best of our knowledge, our Weather2K is the first attempt to tackle weather forecasting task by taking full advantage of the strengths of observation data from ground weather stations.
no code implementations • 30 Jan 2023 • Zhenduo Wang, Yuancheng Tu, Corby Rosset, Nick Craswell, Ming Wu, Qingyao Ai
In this work, we innovatively explore generating clarifying questions in a zero-shot setting to overcome the cold start problem and we propose a constrained clarifying question generation system which uses both question templates and query facets to guide the effective and precise question generation.
no code implementations • 19 Jan 2023 • Shizun Wang, Weihong Zeng, Xu Wang, Hao Yang, Li Chen, Yi Yuan, Yunzhao Zeng, Min Zheng, Chuang Zhang, Ming Wu
To this end, we propose SwiftAvatar, a novel avatar auto-creation framework that is evidently superior to previous works.
2 code implementations • 14 Dec 2022 • Xuandong Zhao, Siqi Ouyang, Zhiguo Yu, Ming Wu, Lei LI
How can we extend a pre-trained model to many language understanding tasks, without labeled or additional unlabeled data?
no code implementations • 25 Nov 2022 • Cheng Lyu, Jiake Xie, Bo Xu, Cheng Lu, Han Huang, Xin Huang, Ming Wu, Chuang Zhang, Yong Tang
Performance of trimap-free image matting methods is limited when trying to decouple the deterministic and undetermined regions, especially in the scenes where foregrounds are semantically ambiguous, chromaless, or high transmittance.
2 code implementations • 5 Sep 2022 • Jinbo Hu, Yin Cao, Ming Wu, Qiuqiang Kong, Feiran Yang, Mark D. Plumbley, Jun Yang
Our system submitted to the DCASE 2022 Task 3 is based on our previous proposed Event-Independent Network V2 (EINV2) with a novel data augmentation method.
1 code implementation • Findings (ACL) 2022 • Xuandong Zhao, Zhiguo Yu, Ming Wu, Lei LI
How to learn highly compact yet effective sentence representation?
3 code implementations • 5 Jan 2022 • Haotian Yan, Chuang Zhang, Ming Wu
In this paper, we succeed in introducing multi-scale representations into semantic segmentation ViT via window attention mechanism and further improves the performance and efficiency.
Ranked #14 on
Semantic Segmentation
on DADA-seg
1 code implementation • 30 Nov 2021 • Shizun Wang, Ming Lu, Kaixin Chen, Jiaming Liu, Xiaoqi Li, Chuang Zhang, Ming Wu
However, existing methods mostly train the DNNs on uniformly sampled LR-HR patch pairs, which makes them fail to fully exploit informative patches within the image.
1 code implementation • ICCV 2021 • Jiaming Liu, Ming Lu, Kaixin Chen, Xiaoqi Li, Shizun Wang, Zhaoqing Wang, Enhua Wu, Yurong Chen, Chuang Zhang, Ming Wu
Internet video delivery has undergone a tremendous explosion of growth over the past few years.
2 code implementations • 27 Apr 2021 • Haotian Yan, Zhe Li, Weijian Li, Changhu Wang, Ming Wu, Chuang Zhang
It is also worth pointing that, given identical strong data augmentations, the performance improvement of ConTNet is more remarkable than that of ResNet.
no code implementations • 1 Jan 2021 • Yaqing Wang, Subhabrata Mukherjee, Haoda Chu, Yuancheng Tu, Ming Wu, Jing Gao, Ahmed Hassan Awadallah
Neural sequence labeling is an important technique employed for many Natural Language Processing (NLP) tasks, such as Named Entity Recognition (NER), slot tagging for dialog systems and semantic parsing.
no code implementations • 1 Jan 2021 • Zhaoqing Wang, Jiaming Liu, Yangyuxuan Kang, Mingming Gong, Chuang Zhang, Ming Lu, Ming Wu
Graph Reasoning has shown great potential recently in modeling long-range dependencies, which are crucial for various computer vision tasks.
no code implementations • 7 Oct 2020 • Yaqing Wang, Subhabrata Mukherjee, Haoda Chu, Yuancheng Tu, Ming Wu, Jing Gao, Ahmed Hassan Awadallah
While self-training serves as an effective mechanism to learn from large amounts of unlabeled data -- meta-learning helps in adaptive sample re-weighting to mitigate error propagation from noisy pseudo-labels.
1 code implementation • ECCV 2020 • Tianyi Wu, Yu Lu, Yu Zhu, Chuang Zhang, Ming Wu, Zhanyu Ma, Guodong Guo
GI unit is further improved by the SC-loss to enhance the semantic representations over the exemplar-based semantic graph.
no code implementations • 15 Mar 2020 • Kaiyan Chen, Ming Wu, Jiaming Liu, Chuang Zhang
To further promote the research of ship detection, we introduced a new fine-grained ship detection datasets, which is named as FGSD.
3 code implementations • 11 Feb 2020 • Dongliang Chang, Yifeng Ding, Jiyang Xie, Ayan Kumar Bhunia, Xiaoxu Li, Zhanyu Ma, Ming Wu, Jun Guo, Yi-Zhe Song
The proposed loss function, termed as mutual-channel loss (MC-Loss), consists of two channel-specific components: a discriminality component and a diversity component.
Ranked #32 on
Fine-Grained Image Classification
on FGVC Aircraft
1 code implementation • 31 Jan 2020 • Yu Lu, Muyan Feng, Ming Wu, Chuang Zhang
Human parsing is an essential branch of semantic segmentation, which is a fine-grained semantic segmentation task to identify the constituent parts of human.
2 code implementations • 22 Nov 2019 • Canran Xu, Ming Wu
Learning representations for feature interactions to model user behaviors is critical for recommendation system and click-trough rate (CTR) predictions.
5 code implementations • CVPR 2018 2018 • Lichen Zhou, Chuang Zhang, Ming Wu
Road extraction is a fundamental task in the field of remote sensing which has been a hot research topic in the past decade.
Ranked #2 on
Road Segmentation
on DeepGlobe
(IoU metric)
no code implementations • 19 Oct 2018 • Lingxiao Ma, Zhi Yang, Youshan Miao, Jilong Xue, Ming Wu, Lidong Zhou, Yafei Dai
This evolution has led to large graph-based irregular and sparse models that go beyond what existing deep learning frameworks are designed for.
no code implementations • 22 May 2018 • Jilong Xue, Youshan Miao, Cheng Chen, Ming Wu, Lintao Zhang, Lidong Zhou
Its computation is typically characterized by a simple tensor data abstraction to model multi-dimensional matrices, a data-flow graph to model computation, and iterative executions with relatively frequent synchronizations, thereby making it substantially different from Map/Reduce style distributed big data computation.