MedGemma is a collection of Gemma 3 variants that are trained for performance on medical text and image comprehension. Developers can use MedGemma to accelerate building healthcare-based AI applications. MedGemma comes in two variants: a 4B multimodal version and a 27B text-only version.
MedGemma 4B utilizes a SigLIP image encoder that has been specifically pre-trained on a variety of de-identified medical data, including chest X-rays, dermatology images, ophthalmology images, and histopathology slides. Its LLM component is trained on a diverse set of medical data, including radiology images, histopathology patches, ophthalmology images, dermatology images, and medical text.
MedGemma variants have been evaluated on a range of clinically relevant benchmarks to illustrate their baseline performance. These include both open benchmark datasets and curated datasets, with a focus on expert human evaluations for tasks. Developers can fine tune MedGemma variants for improved performance. Please read more about our work in our manuscript [link coming] and consult our Intended Use Statement for more details.
-
Read our developer documentation to see the full range of next steps available, including learning more about the model through its model card.
-
Explore this repository, which contains notebooks for using the model.
-
Visit the model on Hugging Face or Model Garden.
We are open to bug reports, pull requests (PR), and other contributions. See CONTRIBUTING and community guidelines for details.
While the model is licensed under the Health AI Developer Foundations License, everything in this repository is licensed under the Apache 2.0 license, see LICENSE.