Megatron-HF Bridge Backend#
AReaL currently supports two bridge backends for MegatronEngine:
mbridge(default)megatron-bridge
Set the backend with:
actor:
megatron:
bridge_type: mbridge
Use
bridge_type=megatron-bridgeto enable the new path.mbridgeis the default choice if this argument is not present
Why this feature exists#
mbridgeis being deprecated and does not provide PEFT/LoRA support.megatron-bridgesupports more/ newer model architectures.megatron-bridgeprovides built-in PEFT/LoRA implementations.
Recommendation#
For new GPU training workflows, prefer
megatron-bridge.Keep
mbridgefor backward compatibility and environments that still depend on it.Prefer
mbridgewhen using disk-based weight broadcast as it has optimized HF load/save path.If you use XCCL for weight broadcast, load/save time is less important.
Current limitation#
Tree-attention training in
MegatronEnginecurrently supports onlymbridge.The
megatron-bridgebackend is not supported in the tree-attention path yet.megatron-bridgedoes support faster/optimized HF model load/save implementations.