Learning Transferable Visual Models From Natural Language Supervision
Paper
•
2103.00020
•
Published
•
19
CLIP Italian is a CLIP-like Model for Italian. The CLIP model (Contrastive Language–Image Pre-training) was developed by researchers at OpenAI and is able to efficiently learn visual concepts from natural language supervision.
We fine-tuned a competitive Italian CLIP model with only ~1.4 million Italian image-text pairs. This model is part of the Flax/Jax Community Week, organized by HuggingFace and TPU usage sponsored by Google.
We considered three main sources of data:
Preprocessing, hardware used, hyperparameters...