From 50948a9ad8fbe006be0f5d145dd1054f7d861c40 Mon Sep 17 00:00:00 2001 From: Terry Kong Date: Fri, 4 Oct 2024 10:04:51 -0700 Subject: [PATCH] fix: serve_reward_model.py no longer errors if pretrained GBS indivisible by parallel state (#333) Signed-off-by: Terry Kong --- examples/nlp/gpt/conf/inference_rm.yaml | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/examples/nlp/gpt/conf/inference_rm.yaml b/examples/nlp/gpt/conf/inference_rm.yaml index 9571ac242..171ef2735 100644 --- a/examples/nlp/gpt/conf/inference_rm.yaml +++ b/examples/nlp/gpt/conf/inference_rm.yaml @@ -49,3 +49,7 @@ model: regression: merge_attributes: False # whether to merge attribute values into a scalar attribute_weights: null # apply these weights to each attributes when merging them into a scalar + + # NOTE: The user does not need to change the global batch size below + # GBS is overridden to 0 to disable checks for compatibility with the megatron-core parallel state + global_batch_size: 0