AReaL currently supports two bridge backends for MegatronEngine:
mbridge(default)megatron-bridge
Set the backend with:
actor:
megatron:
bridge_type: mbridge- Use
bridge_type=megatron-bridgeto enable the new path. mbridgeis the default choice if this argument is not present
mbridgeis being deprecated and does not provide PEFT/LoRA support.megatron-bridgesupports more/ newer model architectures.megatron-bridgeprovides built-in PEFT/LoRA implementations.
- For new GPU training workflows, prefer
megatron-bridge. - Keep
mbridgefor backward compatibility and environments that still depend on it. - Prefer
mbridgewhen using disk-based weight broadcast as it has optimized HF load/save path. - If you use XCCL for weight broadcast, load/save time is less important.
- Tree-attention training in
MegatronEnginecurrently supports onlymbridge. - The
megatron-bridgebackend is not supported in the tree-attention path yet. megatron-bridgedoes support faster/optimized HF model load/save implementations.