-
Notifications
You must be signed in to change notification settings - Fork 281
Insights: pytorch/ao
Overview
-
- 5 Merged pull requests
- 6 Open pull requests
- 0 Closed issues
- 1 New issue
Could not load contribution data
Please try again later
5 Pull requests merged by 5 people
-
add cast config for fp8 enablement
#2328 merged
Jun 10, 2025 -
Fix Per Tensor 3d rehsape
#2293 merged
Jun 9, 2025 -
Update Quantization docs to show newer AOConfigs
#2317 merged
Jun 9, 2025 -
Enhance test_autoquant_compile to support ROCm
#2100 merged
Jun 9, 2025 -
Migrate xnnpack/vulkan/boltnn pt2e from torch.ao to torchao (#11363)
#2302 merged
Jun 9, 2025
6 Pull requests opened by 5 people
-
Replace debug handle with `from_node` to trace operator transformation
#2339 opened
Jun 9, 2025 -
Implemented a new test case for LUT quantization
#2342 opened
Jun 9, 2025 -
Inference tutorial - Part 3 of e2e series [WIP]
#2343 opened
Jun 9, 2025 -
[BE] Rename qparams for tinygemm
#2344 opened
Jun 9, 2025 -
Add inplace quantizer examples
#2345 opened
Jun 10, 2025 -
Add Tutorial on E2E integration into VLLM and minimal Subclass
#2346 opened
Jun 10, 2025
1 Issue opened by 1 person
-
Distributing ao tensor subclasses in .safetensors checkpoints
#2338 opened
Jun 9, 2025
14 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
DUMMY PR: add support for hpu in float8 base and compile test for torch ao
#2326 commented on
Jun 10, 2025 • 14 new comments -
float8 moe training conversion API prototype
#2275 commented on
Jun 10, 2025 • 4 new comments -
ROCm mx-fp8 Gemm
#2066 commented on
Jun 10, 2025 • 2 new comments -
moe quant with dedicated kernels [wip]
#2325 commented on
Jun 9, 2025 • 2 new comments -
Add support for bmm and `to` for fbgemm Tensor
#2337 commented on
Jun 9, 2025 • 2 new comments -
skip quant/dequant decomposed
#2299 commented on
Jun 10, 2025 • 1 new comment -
Add round_scales_to_power_of_2 option for float quantization
#2323 commented on
Jun 9, 2025 • 1 new comment -
[Question] Combining QAT and Sparsity Training
#2310 commented on
Jun 9, 2025 • 0 new comments -
[Windows][build]two Build failure on Windows on latest main branch
#2297 commented on
Jun 9, 2025 • 0 new comments -
int4_weight_only get plain weight are padded
#2249 commented on
Jun 10, 2025 • 0 new comments -
Eval hf models using lm_eval
#2179 commented on
Jun 9, 2025 • 0 new comments -
[WIP] Enable Int4WeightOnlyGPTQQuantizer on Intel GPU.
#2200 commented on
Jun 9, 2025 • 0 new comments -
Fix failing tests on h100
#2231 commented on
Jun 9, 2025 • 0 new comments -
[BE] Make ScalingGranularity module level so it can be rendered in API ref on docsite
#2314 commented on
Jun 9, 2025 • 0 new comments