Skip to content

Conversation

@metascroy
Copy link
Contributor

TorchAO is deprecating AffineQuantizedTensor (AQT), which is what Int8DynamicActivationInt4WeightConfig is based on.

We migrate int8-int4 QAT to use Int8DynamicActivationIntxWeightConfig instead, which does not depend on AQT.

cc @andrewor14

Copy link
Contributor

@andrewor14 andrewor14 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks, cc @danielhanchen

@metascroy
Copy link
Contributor Author

@andrewor14 @danielhanchen I don't have permission to merge

@mmathew23
Copy link
Collaborator

@metascroy I think only issue here is backwards compatibility. torchao==0.11.0 doesn't seem to make available Int8DynamicActivationIntxWeightConfig in torchao.quantization.__init__.

Is it possible to handle that case? Or would you not recommend running with 0.11.0?

@andrewor14
Copy link
Contributor

andrewor14 commented Oct 1, 2025

@mmathew23 I think unsloth QAT requires torchao 0.13.0 actually. This is a new feature added very recently and uses some primitives added in 0.13.0, so I would recommend upgrading to that version instead

@danielhanchen danielhanchen merged commit 64c3330 into unslothai:main Oct 2, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

4 participants