We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent a9a0969 commit ea640a1Copy full SHA for ea640a1
tests/unittest/_torch/modules/test_fused_moe.py
@@ -740,6 +740,7 @@ def per_rank_test_fused_moe_alltoall_fp8_blockwise(job_id):
740
)
741
alltoall_model.to("cuda")
742
alltoall_model.load_weights([weights])
743
+ alltoall_model.post_load_weights()
744
745
# Use DeepGemmFusedMoE as reference
746
ref_model = DeepGemmFusedMoE(
@@ -755,6 +756,7 @@ def per_rank_test_fused_moe_alltoall_fp8_blockwise(job_id):
755
756
757
ref_model.to("cuda")
758
ref_model.load_weights([weights])
759
+ ref_model.post_load_weights()
760
761
# Evaluate the outputs on variant sequence lengths
762
m = MAX_NUM_TOKENS
0 commit comments