p1atdev commited on
Commit
7c54729
·
verified ·
1 Parent(s): fa08d09

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -17
README.md CHANGED
@@ -7,21 +7,5 @@ base_model_relation: quantized
7
 
8
  FP8 quantized version of [AuraFlow v0.3](fal/AuraFlow-v0.3)
9
 
10
- ## Quantization
11
 
12
- ```py
13
- import torch
14
- from huggingface_hub import cached_download
15
- from safetensors.torch import load_file, save_file
16
-
17
- ckpt_path = cached_download(
18
- "https://huggingface.co/fal/AuraFlow-v0.3/resolve/main/aura_flow_0.3.safetensors",
19
- )
20
-
21
- state_dict = load_file(ckpt_path)
22
-
23
- for key, value in state_dict.items():
24
- state_dict[key] = value.to(torch.float8_e4m3fn)
25
-
26
- save_file(state_dict, "./aura_flow_0.3.float8_e4m3fn.safetensors")
27
- ```
 
7
 
8
  FP8 quantized version of [AuraFlow v0.3](fal/AuraFlow-v0.3)
9
 
10
+ Just casted to `torch.float8_e4m3fn` all linear weights of the flow transformer except `t_embedder`, `final_linear`, `modF`.
11