π Registraion | π View Registered Submissions
The ImageCLEFmed-MEDVQA-GI (3rd edition) challenge π¬ focuses on integrating Visual Question Answering (VQA) with synthetic gastrointestinal (GI) data π₯ to enhance diagnostic accuracy πββοΈπ‘ and AI learning algorithms π€.
This year's challenge includes two exciting subtasks π designed to push the boundaries of image analysis πΌοΈ and synthetic medical image generation π§¬, aiming to improve diagnostic processes π¨ and patient outcomes π.
π‘ Goal: This subtask requires participants to develop AI models capable of accurately interpreting and answering clinical questions based on gastrointestinal (GI) images from the Kvasir-VQA dataset. The dataset consists of 6,500 annotated images covering a range of conditions and medical instruments. Questions are categorized into six types: Yes/No, Single-Choice, Multiple-Choice, Color-Related, Location-Related, and Numerical Count, necessitating the processing of both visual and textual information. Model performance will be evaluated using multiple quantitative metrics.
β¨ Focus: Create robust systems that combine image πΌοΈ and text understanding π¨οΈ to assist medical diagnostics π¨.
- π’ How many polyps are in the image?
- β‘ Are there any abnormalities in the image?
- π·οΈ What disease is visible in the image?
ποΈ Goal: Generate synthetic GI images 𧬠that are indistinguishable from real medical images π₯, rich in detail and variability.
π± Why? Provide privacy-preserving alternatives π to real patient data and support diagnostic systems π‘.
Demo with HuggingFace Diffusers
The 2025 dataset ποΈ is an extended version of the HyperKvasir dataset π (datasets.simula.no/hyper-kvasir) and includes:
- π₯ More images (from KVASIR-VQA) than previous years with detailed VQA annotations simulating realistic diagnostic scenarios π
- π― Synthetically generated captions that can be used for image generation task. π οΈ
- π Development Dataset: Kvasir-VQA and captions.
- π Test Dataset: Coming Soon β³ You can split the training dataset for model development now.
- π Metrics: π― Accuracy, π Precision, β»οΈ Recall, and π F1 Score.
- π Evaluation: Based on correctness β and relevance π of answers using the provided questions π¬ and images πΌοΈ.
- π Subjective Evaluation: π©Ί Expert reviewers will assess realism π and diagnostic utility π₯.
- π― Objective Evaluation:
- π FrΓ©chet Inception Distance (FID): Similarity between synthetic and real images.
- ποΈ Structural Similarity Index Measure (SSIM): Resemblance in structure ποΈ.
π View Registered Submissions
We use the medvqa
Python package to validate and submit models to the official system.
The model that needs to be submiited is expected to be in a HuggingFace repository.
pip install -U medvqa
The library is under active development. Always ensure you're using the latest version.
Your HuggingFace repo must include a standalone script named:
submission_task1.py
for Task 1submission_task2.py
for Task 2
Use the provided template script, and make sure to:
- Modify all
TODO
sections - Add required information directly in the script
First make sure your submission script works fine in your working environment and it loads the model correctly from your submission repo and generates outputs in the required format.
python submission_task1.py
Next, you can validate the script to work independently. The .py script should now be in the root of the same HuggingFace repo as your model. You can try this in a new venv:
medvqa validate --competition=gi-2025 --task=1/2 --repo_id=<your_repo_id>
--competition
: Set togi-2025
--task
: Use1
for Task 1 or2
for Task 2--repo_id
: Your HuggingFace model repo ID (e.g.,SushantGautam/XXModelCheckpoint
)
If your code requires extra packages, you must include a requirements.txt
in the root of the repo. The system will install these automatically during validation/submission.
Else you will get package missing errors.
If validation is okey, you can just run:
medvqa validate_and_submit --competition=gi-2025 --task=1/2 --repo_id=<your_repo_id>
This will make a submisision and your username, along with the task and time, should be visible on the portal for it to be considered officially submitted. The submission library will make your Hugging Face repository public but gated, granting the organizers access to your repo. It must remain unchanged at least until the results of the competition are announced. However, you are free to make your model fully public (non-gated).
If you encounter any issues with submission, donβt hesitate to contact us.
- π 20 December 2024: π Registration opens
- π 14 February 2025: π Release of training & validation datasets
- π 9 April 2025: β³ Test datasets released
- π 25 April 2025: πͺ Registration closes
- π 10 May 2025: β²οΈ Run submission deadline
- π 17 May 2025: π Processed results released
- π 30 May 2025: βοΈ Participant papers submission [CEUR-WS]
- π 27 June 2025: π Notification of acceptance
- π 7 July 2025: π¨οΈ Camera-ready paper submission [CEUR-WS]
- ποΈ 9-12 September 2025: π CLEF 2025, Madrid, Spain πͺπΈ
β¨ For any queries, feel free to reach out to our amazing team:
- π¨βπ¬ Steven A. Hicks π§ [email protected]
- π§βπ» Michael A. Riegler π§ [email protected]
- π§βπ¬ Vajira Thambawita π§ [email protected]
- π¨βπ« PΓ₯l Halvorsen π§ [email protected]
- π§βπ Sushant Gautam π§ [email protected]
π Visit: π imageclef.org/2025#registration
π View Registered Submissions: π simulamet-medvqa.hf.space
π₯ Join the challenge, push the boundaries, and make a difference in medical AI! ππ§¬