Skip to content

Conversation

@xiaowangintel
Copy link
Collaborator

@xiaowangintel xiaowangintel commented Oct 27, 2025

The MXFP8/MXFP4/MXFP6 format training and inference is already enabled on CUDA, but not on Intel GPU. This pr use to support mx function of ao on Intel GPU, and enable the corresponding test. Currently, the linear for mx format is only supported on emulate mode on Intel GPU. We are working the scaled_mm on Intel and will enable more GEMM choice after the scaled_mm on xpu is ready.

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 27, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/3248

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

✅ No Failures

As of commit 64f4151 with merge base 03c2d28 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Oct 27, 2025
@xiaowangintel xiaowangintel added ciflow/xpu label used to trigger xpu CI jobs xpu Intel XPU related features labels Oct 27, 2025
@liangan1 liangan1 added the topic: improvement Use this tag if this PR is an improvement (doesn't fit into any of the other categories) label Oct 27, 2025
if torch.xpu.is_available():
devices.append("xpu")


Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggest to use utils function of torchao to get the available device.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The get_available_devices(https://github.com/pytorch/ao/blob/main/torchao/utils.py#L139) interface can get devices. However, the output includes "cpu" which is not what we need.

if torch.xpu.is_available():
devices.append("xpu")


Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggest to use utils function of torchao to get the available device.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ditto

Copy link
Collaborator

@liangan1 liangan1 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

General LGTM.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/xpu label used to trigger xpu CI jobs CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. topic: improvement Use this tag if this PR is an improvement (doesn't fit into any of the other categories) xpu Intel XPU related features

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants