Skip to content

Commit 237a142

Browse files
fix v4 (#1743)
1 parent 6fe20ef commit 237a142

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

_posts/2024-09-25-pytorch-native-architecture-optimization.md renamed to _posts/2024-09-26-pytorch-native-architecture-optimization.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ But also can do things like quantize weights to int4 and the kv cache to int8 to
7272

7373
Post training quantization, especially at less than 4 bit can suffer from serious accuracy degradations. Using [Quantization Aware Training](https://pytorch.org/blog/quantization-aware-training/) (QAT) we’ve managed to recover up to 96% of the accuracy degradation on hellaswag. We’ve integrated this as an end to end recipe in torchtune with a minimal [tutorial](https://github.com/pytorch/ao/tree/main/torchao/quantization/prototype/qat)
7474

75-
![](/assets/assets/Figure_3.png){:style="width:100%"}
75+
![](/assets/images/Figure_3.png){:style="width:100%"}
7676

7777
# Training
7878

0 commit comments

Comments
 (0)