Effect of Pre-Training Scale on Intra- and Inter-Domain Full and Few-Shot Transfer Learning for Natural and Medical X-Ray Chest Images

31 May 2021  ·  Mehdi Cherti, Jenia Jitsev ·

Increasing model, data and compute budget scale in the pre-training has been shown to strongly improve model generalization and transfer learning in vast line of work done in language modeling and natural image recognition. However, most studies on the positive effect of larger scale were done in scope of in-domain setting, with source and target data being in close proximity. To study effect of larger scale for both in-domain and out-of-domain setting when performing full and few-shot transfer, we combine here for the first time large, openly available medical X-Ray chest imaging datasets to reach a scale for medical imaging domain comparable to ImageNet-1k, routinely used for pre-training in natural image domain. We then conduct supervised pre-training, while varying network size and source data scale and domain, being either large natural (ImageNet-1k/21k) or large medical chest X-Ray datasets, and transfer pre-trained models to different natural or medical targets. We observe strong improvement due to larger pre-training scale for intra-domain natural-natural and medical-medical transfer. For inter-domain natural-medical transfer, we find improvements due to larger pre-training scale on larger X-Ray targets in full shot regime, while for smaller targets and for few-shot regime the improvement is not visible. Remarkably, large networks pre-trained on very large natural ImageNet-21k are as good or better than networks pre-trained on largest available medical X-Ray data when performing transfer to large X-Ray targets. We conclude that substantially increasing model and generic, medical domain-agnostic natural image source data scale in the pre-training can enable high quality out-of-domain transfer to medical domain specific targets, removing dependency on large medical domain-specific source data often not available in the practice.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Image Classification CIFAR-10 ResNet-152x4-AGC (ImageNet-21K) Percentage correct 97.82 # 63
Image Classification CIFAR-10 ResNet-50x1-ACG (ImageNet-21K) Percentage correct 95.78 # 116
Image Classification CIFAR-100 ResNet-152x4-AGC (ImageNet-21K) Percentage correct 88.54 # 37
Image Classification Flowers-102 ResNet-152x4-AGC (ImageNet-21K) Accuracy 99.49 # 10
Image Classification Flowers-102 ResNet-50x1-ACG (ImageNet-21K) Accuracy 98.21 # 26
Image Classification Oxford-IIIT Pet Dataset ResNet-152x4-AGC (ImageNet-21K) Accuracy 93.21 # 2

Methods