Skip to content

Conversation

@Fridah-nv
Copy link
Contributor

What does this PR do?

Type of change: ?

Overview: ?

Usage

# Add a code snippet demonstrating how to use this

Testing

Before your PR is "Ready for review"

  • Make sure you read and follow Contributor guidelines and your commits are signed.
  • Is this change backward compatible?: Yes/No
  • Did you write any new necessary tests?: Yes/No
  • Did you add or update any necessary documentation?: Yes/No
  • Did you update Changelog?: Yes/No

Additional Information

@Fridah-nv Fridah-nv self-assigned this Nov 27, 2025
@copy-pr-bot
Copy link

copy-pr-bot bot commented Nov 27, 2025

Auto-sync is disabled for draft pull requests in this repository. Workflows must be run manually.

Contributors can view more details about this message here.

@codecov
Copy link

codecov bot commented Nov 27, 2025

Codecov Report

❌ Patch coverage is 71.42857% with 2 lines in your changes missing coverage. Please review.
✅ Project coverage is 74.65%. Comparing base (f06c3f9) to head (c025df7).
⚠️ Report is 5 commits behind head on main.

Files with missing lines Patch % Lines
.../torch/quantization/nn/modules/tensor_quantizer.py 60.00% 2 Missing ⚠️
Additional details and impacted files
@@            Coverage Diff             @@
##             main     #613      +/-   ##
==========================================
+ Coverage   74.57%   74.65%   +0.08%     
==========================================
  Files         183      183              
  Lines       18412    18546     +134     
==========================================
+ Hits        13730    13845     +115     
- Misses       4682     4701      +19     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.



@triton.jit
def blockwise_fp4_fake_quant_kernel(
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
def blockwise_fp4_fake_quant_kernel(
def static_blockwise_fp4_fake_quant_kernel(

disable_calib(quantizer),
enable_fake_quant(quantizer),
):
quantizer._keep_shape = True
Copy link
Contributor

@realAsma realAsma Dec 3, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

line 233:

if is_nvfp4_static: # static per-block
amax = scaled_e4m3_impl(amax, amax.amax()) # fP8 quantization

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants