CC3M更简洁,更适合微调,但也可以与CC12M一起用于预训练,论文说明了这一点。巧合的是,它们的交集并不为零--大约有 63K 个 URL。 联系我们 如果您有上述常见问题解答中没有的问题,或者您想分享反馈或报告问题,请发送电子邮件至 conceptual-captions@google.com。
We take a step further in pushing the limits of vision-and-language pre-training data by relaxing the data collection pipeline used in Conceptual Captions 3M (CC3M) [70] and introduce the Conceptual 12M (CC12M), a dataset with 12 million image-text pairs specifically meant to be used for...
Add a description, image, and links to the conceptual-captions topic page so that developers can more easily learn about it. Curate this topic Add this topic to your repo To associate your repository with the conceptual-captions topic, visit your repo's landing page and select "manage to...