Med-PRM: Medical Reasoning Models with Stepwise, Guideline-verified Process Rewards


Loading...

Date

2025-11

Publication Type

Conference Paper

ETH Bibliography

yes

Citations

Altmetric

Data

Abstract

Large language models have shown promise in clinical decision making, but current approaches struggle to localize and correct errors at specific steps of the reasoning process. This limitation is critical in medicine, where identifying and addressing reasoning errors is essential for accurate diagnosis and effective patient care. We introduce Med-PRM, a process reward modeling framework that leverages retrieval-augmented generation to verify each reasoning step against established medical knowledge bases. By verifying intermediate reasoning steps with evidence retrieved from clinical guidelines and literature, our model can precisely assess the reasoning quality in a fine-grained manner. Evaluations on five medical QA benchmarks and two open-ended diagnostic tasks demonstrate that Med-PRM achieves state-of-the-art performance, with improving the performance of base models by up to 13.50% using Med-PRM. Moreover, we demonstrate the generality of Med-PRM by integrating it in a plug-and-play fashion with strong policy models such as Meerkat, achieving over 80% accuracy on MedQA for the first time using small-scale models of 8 billion parameters.

Publication status

published

Book title

Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing

Journal / series

Volume

Pages / Article No.

16554 - 16571

Publisher

Association for Computational Linguistics

Event

30th Conference on Empirical Methods in Natural Language Processing (EMNLP 2025)

Edition / version

Methods

Software

Geographic location

Date collected

Date created

Subject

Organisational unit

09847 - Moor, Michael / Moor, Michael check_circle

Notes

Funding

Related publications and datasets