-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Inference] Qwen2 support fp8 inference #8954
Conversation
Thanks for your contribution! |
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## develop #8954 +/- ##
===========================================
+ Coverage 53.34% 53.44% +0.09%
===========================================
Files 652 652
Lines 105484 105188 -296
===========================================
- Hits 56270 56214 -56
+ Misses 49214 48974 -240 ☔ View full report in Codecov by Sentry. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* qwen2 fp8 * fp8 check * fp8 cutlass * int8 cachekv * a8w8c8_fp8
PR types
New features
PR changes
Others
Description
qwen2 支持a8w8_fp8和a8w8c8_fp8推理