Replies: 1 comment 1 reply
-
|
float W8 doesn't work for full finetuning either. If it doesn't give you an error, it will fail to learn anything. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
When performing full fine-tuning on z-image models, selecting specific Transformer DataType options related to quantization triggers a runtime error.
RuntimeError: only Tensors of floating point dtype can require gradientsThe error occurs when setting transformer DataType to any of the following quantization modes:
int W8A8
float W8A8
GGUF
Notably, the float (W8) mode works normally during full fine-tuning – is there any fundamental difference between float (W8) and the other modes (float W8A8/int W8A8/GGUF) that explains this compatibility?
Beta Was this translation helpful? Give feedback.
All reactions