🛰️ Official repository of paper "RemoteCLIP: A Vision Language Foundation Model for Remote Sensing" (IEEE TGRS)
-
Updated
Jun 27, 2024 - Jupyter Notebook
🛰️ Official repository of paper "RemoteCLIP: A Vision Language Foundation Model for Remote Sensing" (IEEE TGRS)
An extremely simple method for validation-free efficient adaptation of CLIP-like VLMs that is robust to the learning rate.
[ICLR2025] Detecting Backdoor Samples in Contrastive Language Image Pretraining
Use CLIP to create matching texts + embeddings for given images; useful for XAI, adversarial training
Simple data and training pipeline for class-incremental method 😄
RadCLIP is a foundation model for radiologic imaging that leverages a Vision–Language Pre-training (VLP) framework to align 2D/3D radiologic images with their textual descriptions, improving diagnostic accuracy and efficiency in clinical workflows.
Theory, Experiments, and Dataset for our newly proposed Deep Learning method for LLM-driven Cycle Consistency and Semantics Aware Self-Supervised Framework for Unpaired LDR ↔ HDR Image Translation
This application fine-tunes the CLIP model on the Flickr8k dataset to align image and text embeddings for image-caption matching. The goal is to enhance multimodal understanding and retrieval performance using a custom captioning dataset.
CLIP-interrogator InvokeAI node
Visual and Vision-Language Representation Pre-Training with Contrastive Learning
Attempts to improve CLIP via different optimizers and loss functions
PLoP applied to CLIP
Train a vision language model from scratch used for chart interpretation in data analysis
University project based on implementing a Test Time Adaptation (TTA) solution for image classifiers. University of Trento (Italy)
Add a description, image, and links to the contrastive-language-image-pretraining topic page so that developers can more easily learn about it.
To associate your repository with the contrastive-language-image-pretraining topic, visit your repo's landing page and select "manage topics."