Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[AutoParallel] Support qwen for auto_parallel #8312

Open
wants to merge 19 commits into
base: develop
Choose a base branch
from

Conversation

GhostScreaming
Copy link
Contributor

@GhostScreaming GhostScreaming commented Apr 23, 2024

PR types

New features

PR changes

Models

Description

Support qwen for auto_parallel.

1. 在QWen上验证自动并行架构

  • 对比动态图半自动并行和动态图手动并行,收敛和精度结果符合预期,涉及以下策略验证
    • dp2mp2pp2
    • dp2mp2pp2 + amp
    • dp2mp2pp2 + 动转静 + amp

2. 本 PR 代码包含最终模型代码、示例脚本、单测。验证结果如下

  • dp2mp2pp2 FP32 Loss Diff

- dp2mp2pp2 BF16 Loss Diff

- dp2mp2pp2 动转静 FP32 Loss Diff

- dp2mp2pp2 动转静 BF16 Loss Diff

Copy link

paddle-bot bot commented Apr 23, 2024

Thanks for your contribution!

Copy link

codecov bot commented Apr 23, 2024

Codecov Report

Attention: Patch coverage is 16.56051% with 393 lines in your changes are missing coverage. Please review.

Project coverage is 55.18%. Comparing base (1ffa290) to head (1799dea).
Report is 6 commits behind head on develop.

Files Patch % Lines
paddlenlp/transformers/qwen/modeling_3D_auto.py 16.38% 393 Missing ⚠️
Additional details and impacted files
@@             Coverage Diff             @@
##           develop    #8312      +/-   ##
===========================================
- Coverage    55.25%   55.18%   -0.07%     
===========================================
  Files          613      614       +1     
  Lines        95626    96326     +700     
===========================================
+ Hits         52837    53161     +324     
- Misses       42789    43165     +376     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@liym27
Copy link
Contributor

liym27 commented May 14, 2024

LGTM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants