Skip to content

Update test_gpt_pp baselines for cross_entropy precision alignment#703

Open
zrr1999 wants to merge 3 commits intoPaddlePaddle:developfrom
zrr1999:paa/update-gpt-pp-baseline-develop
Open

Update test_gpt_pp baselines for cross_entropy precision alignment#703
zrr1999 wants to merge 3 commits intoPaddlePaddle:developfrom
zrr1999:paa/update-gpt-pp-baseline-develop

Conversation

@zrr1999
Copy link
Copy Markdown
Member

@zrr1999 zrr1999 commented Mar 31, 2026

Summary

Context

Paddle PR #78044 aligns cross_entropy precision with PyTorch under FLAGS_use_accuracy_compatible_kernel. The grad kernel change (using AccT instead of native T arithmetic) affects gradient computation even in the default path, causing the test_gpt_pp multi-card test to fail with mismatched baseline checksums.

New baseline values are taken from the CI run output of PR #78044.

Test plan

  • Fleet Unit test (multi-card) should pass with updated baselines when paired with Paddle PR #78044

🤖 Generated with Claude Code

zrr1999 and others added 2 commits March 31, 2026 10:55
…gnment

Update MD5 baseline checksums in test_gpt_pp.py to match the new gradient
values produced by Paddle PR #78044, which aligns cross_entropy precision
with PyTorch under FLAGS_use_accuracy_compatible_kernel.

The cross_entropy grad kernel now uses AccT (MPType) arithmetic instead of
native T arithmetic, which changes intermediate gradient values and
propagates through the pipeline parallel backward pass.

Loss MD5 remains unchanged (bce3fed95247f1b7a165e32b33d6fca7).

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Re-enable TestPP that was temporarily skipped by PR PaddlePaddle#698 while
waiting for Paddle PR #78013 to merge.

Co-authored-by: Copilot <223556219+Copilot@users.noreply.github.com>
@zrr1999 zrr1999 force-pushed the paa/update-gpt-pp-baseline-develop branch from f3fd895 to 4371432 Compare March 31, 2026 11:30
@zrr1999
Copy link
Copy Markdown
Member Author

zrr1999 commented Apr 1, 2026

/re-run all-failed

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant