Scaling up vision-language pre-training for image captioning

X Hu, Z Gan, J Wang, Z Yang, Z Liu… - Proceedings of the …, 2022 - openaccess.thecvf.com
In recent years, we have witnessed significant performance boost in the image captioning
task based on vision-language pre-training (VLP). Scale is believed to be an important factor …

Scaling Up Vision-Language Pretraining for Image Captioning

X Hu, Z Gan, J Wang, Z Yang, Z Liu… - 2022 IEEE/CVF …, 2022 - ieeexplore.ieee.org
In recent years, we have witnessed significant performance boost in the image captioning
task based on vision-language pre-training (VLP). Scale is believed to be an important factor …

Scaling Up Vision-Language Pre-training for Image Captioning

X Hu, Z Gan, J Wang, Z Yang, Z Liu, Y Lu… - arXiv preprint arXiv …, 2021 - arxiv.org
In recent years, we have witnessed significant performance boost in the image captioning
task based on vision-language pre-training (VLP). Scale is believed to be an important factor …

Scaling Up Vision-Language Pre-training for Image Captioning

X Hu, Z Gan, J Wang, Z Yang, Z Liu, Y Lu… - arXiv e …, 2021 - ui.adsabs.harvard.edu
In recent years, we have witnessed significant performance boost in the image captioning
task based on vision-language pre-training (VLP). Scale is believed to be an important factor …

Scaling Up Vision-Language Pretraining for Image Captioning

X Hu, Z Gan, J Wang, Z Yang, Z Liu, Y Lu… - 2022 IEEE/CVF …, 2022 - computer.org
In recent years, we have witnessed significant performance boost in the image captioning
task based on vision-language pre-training (VLP). Scale is believed to be an important factor …