Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model Built on a 1/32 Activation-Ratio MoE Architecture

“`html Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model MarkTechPost What Is AntAngelMed? AntAngelMed…

By AI Maestro May 12, 2026 2 min read
Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model Built on a 1/32 Activation-Ratio MoE Architecture

“`html




Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model

Meet AntAngelMed: A 103B-Parameter Open-Source Medical Language Model

MarkTechPost

What Is AntAngelMed?

AntAngelMed is a large-scale medical-domain language model with 103 billion total parameters. However, it activates only 6.1 billion parameters during inference through a Mixture-of-Experts (MoE) architecture with a 1/32 activation ratio. This design allows for efficient computation while maintaining strong knowledge capacity.

Training Pipeline

The model undergoes a three-stage training process:

  • Stage 01: Continual Pre-Training – The model is pre-trained on large-scale medical corpora including encyclopedias, web text, and academic publications. This phase builds the foundational general reasoning capability.
  • Stage 02: Supervised Fine-Tuning (SFT) – The model is fine-tuned with a multi-source instruction dataset that includes both general tasks like math and programming as well as medical scenarios such as doctor–patient Q&A, diagnostic reasoning, and safety/ethics cases.
  • Stage 03: Reinforcement Learning via GRPO – The model is further refined using the Group Relative Policy Optimization (GRPO) algorithm. This stage uses task-specific reward models to encourage the model to behave in ways that are empathetic, structurally clear, safe, and evidence-based.

Inference Performance

On H20 hardware, AntAngelMed achieves a throughput of over 200 tokens per second. This is approximately three times faster than a comparable 36 billion parameter dense model. When combined with the EAGLE3 speculative decoding optimization using FP8 quantization, the inference throughput improves significantly across various benchmarks:

  • HumanEval: +71%
  • GSM8K: +45%
  • Math-500: +94%

The model supports a context length of 128K, which is sufficient for handling full clinical documents and extended patient histories.

Benchmark Results

  • HealthBench: AntAngelMed ranks first among all open-source models, outperforming proprietary top models particularly in the HealthBench-Hard subset.
  • MedAIBench: It ranks at the top level with strong performance across medical knowledge Q&A and medical ethics/safety categories.
  • MedBench: AntAngelMed is the first-ranking model overall, demonstrating its comprehensive capabilities in various areas of healthcare language models.

MarkTechPost’s Visual Explainer




“`

This HTML document contains the rewritten text within appropriate HTML tags and structure. It includes a basic CSS link for styling purposes and JavaScript links that would be used to enhance functionality, though they are not present in this snippet.


Originally published at marktechpost.com. Curated by AI Maestro.

Stay ahead of AI. Get the most important stories delivered to your inbox — no spam, no noise.

Name
Scroll to Top