No Metrics Are Perfect: Adversarial Reward Learning for Visual Storytelling

Though impressive results have been achieved in visual captioning, the task of generating abstract stories from photo streams is still a little-tapped problem. Different from captions, stories have more expressive language styles and contain many imaginary concepts that do not appear in the images. Thus it poses challenges to behavioral cloning algorithms. Furthermore, due to the limitations of automatic metrics on evaluating story quality, reinforcement learning methods with hand-crafted rewards also face difficulties in gaining an overall performance boost. Therefore, we propose an Adversarial REward Learning (AREL) framework to learn an implicit reward function from human demonstrations, and then optimize policy search with the learned reward function. Though automatic eval- uation indicates slight performance boost over state-of-the-art (SOTA) methods in cloning expert behaviors, human evaluation shows that our approach achieves significant improvement in generating more human-like stories than SOTA systems.

PDF Abstract ACL 2018 PDF ACL 2018 Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Visual Storytelling VIST GAN BLEU-1 62.8 # 11
BLEU-2 38.8 # 8
BLEU-3 23.0 # 11
BLEU-4 14 # 15
METEOR 35 # 21
CIDEr 9 # 16
ROUGE-L 29.5 # 22
Visual Storytelling VIST XE-ss BLEU-1 62.3 # 12
BLEU-2 38.2 # 10
BLEU-3 22.5 # 12
BLEU-4 13.7 # 17
METEOR 34.8 # 23
CIDEr 8.7 # 18
ROUGE-L 29.7 # 19
Visual Storytelling VIST AREL-t-100 BLEU-1 63.8 # 8
BLEU-2 39.1 # 7
BLEU-3 23.2 # 9
BLEU-4 14.1 # 13
METEOR 35 # 21
CIDEr 9.4 # 13
ROUGE-L 29.5 # 22

Methods


No methods listed for this paper. Add relevant methods here