Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CLIP’s Impressive Generalization – Any Future Updates? #489

Open
arifur-rahman-ar opened this issue Mar 12, 2025 · 0 comments
Open

CLIP’s Impressive Generalization – Any Future Updates? #489

arifur-rahman-ar opened this issue Mar 12, 2025 · 0 comments

Comments

@arifur-rahman-ar
Copy link

After reading the CLIP paper, I’m highly impressed by its ability to perform zero-shot learning and generalize across image-text tasks without task-specific fine-tuning. The contrastive learning approach, combined with large-scale internet pretraining, allows CLIP to match ResNet50 on ImageNet without labeled examples, which is a significant achievement.
However, I have a few questions regarding future improvements:

1.Model Variants: Are there any plans to release additional CLIP model variants with different architectures or training strategies?
2.Fine-Tuning Support: While CLIP excels at zero-shot learning, is there an official recommendation or upcoming support for fine-tuning it on specific datasets?
3.Performance on Complex Queries: Have there been any internal evaluations or planned improvements for handling more complex, multi-part queries?

Looking forward to any insights on these points. Thanks for the amazing work!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant