Update README.md
Browse files
README.md
CHANGED
|
@@ -7,21 +7,5 @@ base_model_relation: quantized
|
|
| 7 |
|
| 8 |
FP8 quantized version of [AuraFlow v0.3](fal/AuraFlow-v0.3)
|
| 9 |
|
| 10 |
-
|
| 11 |
|
| 12 |
-
```py
|
| 13 |
-
import torch
|
| 14 |
-
from huggingface_hub import cached_download
|
| 15 |
-
from safetensors.torch import load_file, save_file
|
| 16 |
-
|
| 17 |
-
ckpt_path = cached_download(
|
| 18 |
-
"https://huggingface.co/fal/AuraFlow-v0.3/resolve/main/aura_flow_0.3.safetensors",
|
| 19 |
-
)
|
| 20 |
-
|
| 21 |
-
state_dict = load_file(ckpt_path)
|
| 22 |
-
|
| 23 |
-
for key, value in state_dict.items():
|
| 24 |
-
state_dict[key] = value.to(torch.float8_e4m3fn)
|
| 25 |
-
|
| 26 |
-
save_file(state_dict, "./aura_flow_0.3.float8_e4m3fn.safetensors")
|
| 27 |
-
```
|
|
|
|
| 7 |
|
| 8 |
FP8 quantized version of [AuraFlow v0.3](fal/AuraFlow-v0.3)
|
| 9 |
|
| 10 |
+
Just casted to `torch.float8_e4m3fn` all linear weights of the flow transformer except `t_embedder`, `final_linear`, `modF`.
|
| 11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|