Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/4191
Note: Links to docs will display an error until the docs builds have been completed. ❌ 8 New FailuresAs of commit 950b4fb with merge base 4611835 ( NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
| if device == "cuda": | ||
| torch.cuda.empty_cache() | ||
| elif device == "xpu": | ||
| torch.xpu.empty_cache() |
There was a problem hiding this comment.
suggest to use the torch.acc API
| # Save model to generated output directory | ||
| print(f"Saving model to {output_dir}...") | ||
| tokenizer.save_pretrained(output_dir) | ||
| print("model:", model) |
There was a problem hiding this comment.
| print("model:", model) |
|
|
||
| def main(): | ||
| args = parse_args() | ||
| device = _get_device(args.device) |
There was a problem hiding this comment.
suggest to use torch.acc.get_current_xxx to get the current device information instead of add new params --device. it will change the user behevior
|
|
||
| @skip_if_lt_x_gpu(2) | ||
| @unittest.skipIf(not torch.accelerator.is_available(), "Need GPU available") | ||
| @unittest.skipIf(torch.xpu.is_available(), "XPU enablement in progress") |
There was a problem hiding this comment.
pls skip not related UT in another PR.
|
this feature is not complete yet, @xiaowangintel @liangan1 can you restrict the contribution only to stable features for now? they are in https://github.com/pytorch/ao/tree/main/torchao/quantization/quantize_ |
Sure. Thanks for your suggestion. We will focus on these features now. |
Summary
This PR enables GPTQ support on Intel GPU.
Previously, GPTQ workflows in torchao were primarily validated on CUDA. This PR extends support to XPU.