Skip to content

πŸ”¬Challenge on VQA and synthetic GI images for enhanced diagnostic AI | 🌐 imageclef.org/2025/medical/vqa | πŸ‡ͺπŸ‡Έ CLEF 2025, Madrid

Notifications You must be signed in to change notification settings

simula/ImageCLEFmed-MEDVQA-GI-2025

Repository files navigation

🌟 ImageCLEFmed-MEDVQA-GI-2025 🌟

πŸ“ Registraion | πŸ“‹ View Registered Submissions

The ImageCLEFmed-MEDVQA-GI (3rd edition) challenge πŸ”¬ focuses on integrating Visual Question Answering (VQA) with synthetic gastrointestinal (GI) data πŸ₯ to enhance diagnostic accuracy πŸƒβ€β™‚οΈπŸ’‘ and AI learning algorithms πŸ€–.

This year's challenge includes two exciting subtasks πŸš€ designed to push the boundaries of image analysis πŸ–ΌοΈ and synthetic medical image generation 🧬, aiming to improve diagnostic processes 🏨 and patient outcomes πŸ’–.


🎯 Task Descriptions

πŸ” Subtask 1: Algorithm Development for Question Interpretation and Response

πŸ’‘ Goal: This subtask requires participants to develop AI models capable of accurately interpreting and answering clinical questions based on gastrointestinal (GI) images from the Kvasir-VQA dataset. The dataset consists of 6,500 annotated images covering a range of conditions and medical instruments. Questions are categorized into six types: Yes/No, Single-Choice, Multiple-Choice, Color-Related, Location-Related, and Numerical Count, necessitating the processing of both visual and textual information. Model performance will be evaluated using multiple quantitative metrics.

✨ Focus: Create robust systems that combine image πŸ–ΌοΈ and text understanding πŸ—¨οΈ to assist medical diagnostics 🏨.

πŸ’¬ Example Questions:

  • πŸ”’ How many polyps are in the image?
  • ⚑ Are there any abnormalities in the image?
  • 🏷️ What disease is visible in the image?

πŸ’₯ Example Training Notebook:

Demo with HuggingFace Trainer Open In Colab

Demo with SWIFT CLI Open In Colab


🎨 Subtask 2: Creation of High-Fidelity Synthetic GI Images

πŸ–ŒοΈ Goal: Generate synthetic GI images 🧬 that are indistinguishable from real medical images πŸ₯, rich in detail and variability.

🌱 Why? Provide privacy-preserving alternatives πŸ”’ to real patient data and support diagnostic systems πŸ’‘.

πŸ’₯ Example Training Notebook:

Demo with HuggingFace Diffusers Open In Colab


πŸ“‚ Data

The 2025 dataset πŸ—ƒοΈ is an extended version of the HyperKvasir dataset πŸ”— (datasets.simula.no/hyper-kvasir) and includes:

πŸ“₯ Datasets

  • πŸƒ Development Dataset: Kvasir-VQA and captions.
  • πŸ•‘ Test Dataset: Coming Soon ⏳ You can split the training dataset for model development now.

πŸ§ͺ Evaluation Methodology

πŸƒ Subtask 1: Question Interpretation and Response

  • πŸ“Š Metrics: 🎯 Accuracy, πŸ” Precision, ♻️ Recall, and πŸ† F1 Score.
  • πŸ“œ Evaluation: Based on correctness βœ… and relevance πŸ“ of answers using the provided questions πŸ’¬ and images πŸ–ΌοΈ.

πŸ–ΌοΈ Subtask 2: Synthetic Image Quality

  • πŸ‘€ Subjective Evaluation: 🩺 Expert reviewers will assess realism 🌟 and diagnostic utility πŸ₯.
  • 🎯 Objective Evaluation:
    • πŸ“‰ FrΓ©chet Inception Distance (FID): Similarity between synthetic and real images.
    • πŸ—οΈ Structural Similarity Index Measure (SSIM): Resemblance in structure πŸ›οΈ.

πŸ† Submission System

πŸš€ View Registered Submissions

We use the medvqa Python package to validate and submit models to the official system. The model that needs to be submiited is expected to be in a HuggingFace repository.

πŸ“¦ Installation

pip install -U medvqa

The library is under active development. Always ensure you're using the latest version.

Your HuggingFace repo must include a standalone script named:

Use the provided template script, and make sure to:

  • Modify all TODO sections
  • Add required information directly in the script

βœ… Validate Before Submitting

First make sure your submission script works fine in your working environment and it loads the model correctly from your submission repo and generates outputs in the required format.

python submission_task1.py

Next, you can validate the script to work independently. The .py script should now be in the root of the same HuggingFace repo as your model. You can try this in a new venv:

medvqa validate --competition=gi-2025 --task=1/2 --repo_id=<your_repo_id>
  • --competition: Set to gi-2025
  • --task: Use 1 for Task 1 or 2 for Task 2
  • --repo_id: Your HuggingFace model repo ID (e.g., SushantGautam/XXModelCheckpoint)

πŸ“„ Additional Dependencies

If your code requires extra packages, you must include a requirements.txt in the root of the repo. The system will install these automatically during validation/submission. Else you will get package missing errors.

πŸš€ Submission Command

If validation is okey, you can just run:

medvqa validate_and_submit --competition=gi-2025 --task=1/2 --repo_id=<your_repo_id>

This will make a submisision and your username, along with the task and time, should be visible on the portal for it to be considered officially submitted. The submission library will make your Hugging Face repository public but gated, granting the organizers access to your repo. It must remain unchanged at least until the results of the competition are announced. However, you are free to make your model fully public (non-gated).

If you encounter any issues with submission, don’t hesitate to contact us.


πŸ—“οΈ Preliminary Schedule

  • πŸ“… 20 December 2024: πŸ“ Registration opens
  • πŸ“… 14 February 2025: πŸƒ Release of training & validation datasets
  • πŸ“… 9 April 2025: ⏳ Test datasets released
  • πŸ“… 25 April 2025: πŸšͺ Registration closes
  • πŸ“… 10 May 2025: ⏲️ Run submission deadline
  • πŸ“… 17 May 2025: πŸ† Processed results released
  • πŸ“… 30 May 2025: ✍️ Participant papers submission [CEUR-WS]
  • πŸ“… 27 June 2025: πŸ’Œ Notification of acceptance
  • πŸ“… 7 July 2025: πŸ–¨οΈ Camera-ready paper submission [CEUR-WS]
  • πŸ›οΈ 9-12 September 2025: 🌍 CLEF 2025, Madrid, Spain πŸ‡ͺπŸ‡Έ

πŸ’Ό Organizers

✨ For any queries, feel free to reach out to our amazing team:


πŸ”— For More Details & Registration

πŸ“ Visit: πŸ‘‰ imageclef.org/2025#registration

πŸ“‹ View Registered Submissions: πŸ‘‰ simulamet-medvqa.hf.space

πŸ’₯ Join the challenge, push the boundaries, and make a difference in medical AI! πŸš€πŸ§¬

About

πŸ”¬Challenge on VQA and synthetic GI images for enhanced diagnostic AI | 🌐 imageclef.org/2025/medical/vqa | πŸ‡ͺπŸ‡Έ CLEF 2025, Madrid

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •