-
Notifications
You must be signed in to change notification settings - Fork 216
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MX: move block_size and elem_dtype into MXLinearConfig #1689
Conversation
Stack from ghstack (oldest at bottom): |
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/1689
Note: Links to docs will display an error until the docs builds have been completed. This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: ghstack-source-id: 92913315b65e8b19a55bab3a51c54ed56c1bb4f4 ghstack-comment-id: 2649054194 Pull Request resolved: #1689
# TODO(future PR): refactor to make this cleaner | ||
elem_dtype_weight_override: Optional[Any] = None | ||
elem_dtype_grad_output_override: Optional[Any] = None | ||
|
||
# If True, uses a custom triton kernel for fp4 dequantize | ||
use_fp4_custom_triton_dequant_kernel: bool = False |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
you think that we will want to keep this public?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
unlikely, but IMO we can punt that until later
Summary:
Moves
block_size
andelem_dtype
intoMXLinearConfig
and updates all callsites.Before
After
Test Plan: CI
Reviewers:
Subscribers:
Tasks:
Tags: