Release Notes

 

04-06-2025

Welcome to John Snow Labs Medicall LLMs on premise deployments Documentation and Updates Hub!

We are excited to announce the launch of the on premise deplyment of our Medical LLM models page, a centralized repository for all the latest features, enhancements, and resolutions of known issues within the. This dedicated space is designed to keep users informed of the most recent developments, enabling seamless testing and facilitating the provision of valuable feedback. Our commitment is to ensure that users have immediate access to the latest information, empowering them to leverage the full capabilities of out Medical LLM models effectively. Stay updated with us as we continue to improve and expand the functionalities of our Medical LLMsto meet and exceed your expectations.

Supported Medical LLM Models

Model Name Parameters Recommended GPU Memory Max Sequence Length Model Size Max KV-Cache Tensor Parallel Sizes
Medical-LLM-7B 7B ~25 GB 16K 14 GB 11 GB 1, 2, 4
Medical-LLM-10B 10B ~35 GB 32K 19 GB 15 GB 1, 2, 4
Medical-LLM-14B 14B ~40 GB 16K 27 GB 13 GB 1, 2
Medical-LLM-24B 24B ~69 GB 32K 44 GB 25 GB 1, 2, 4, 8
Medical-LLM-Small 14B ~58 GB 32K 28 GB 30 GB 1, 2, 4, 8
Medical-LLM-Medium 70B ~452 GB 128K 131 GB 320 GB 4, 8
Medical-Reasoning-LLM-14B 14B ~58 GB 32K 28 GB 30 GB 1, 2, 4, 8
Medical-Reasoning-LLM-32B 32B ~222 GB 128K 61 GB 160 GB 2, 4, 8
Last updated