Exploring Train and Test-Time Augmentations for Audio-Language Learning

31 Oct 2022  ·  Eungbeom Kim, Jinhee Kim, Yoori Oh, KyungSu Kim, Minju Park, Jaeheon Sim, Jinwoo Lee, Kyogu Lee ·

In this paper, we aim to unveil the impact of data augmentation in audio-language multi-modal learning, which has not been explored despite its importance. We explore various augmentation methods at not only train-time but also test-time and find out that proper data augmentation can lead to substantial improvements. Specifically, applying our proposed audio-language paired augmentation PairMix, which is the first multi-modal audio-language augmentation method, outperforms the baselines for both automated audio captioning and audio-text retrieval tasks. To fully take advantage of data augmentation, we also present multi-level test-time augmentation (Multi-TTA) for the test-time. We successfully incorporate the two proposed methods and uni-modal augmentations and achieve 47.5 SPIDEr on audio captioning, which is an 18.2% relative increase over the baseline. In audio-text retrieval, the proposed methods also show an improvement in performance as well.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Audio captioning AudioCaps AL-MixGen CIDEr 0.755 # 7
SPIDEr 0.466 # 6
SPICE 0.177 # 6
Text to Audio Retrieval AudioCaps AL-MixGen + Multi-TTA R@1 34.7 # 6
R@10 83.3 # 3
Audio to Text Retrieval AudioCaps AL-MixGen + Multi-TTA R@1 40.0 # 2
R@10 87.0 # 2

Methods


No methods listed for this paper. Add relevant methods here