Can someone make quantized gguf or tell me how to do it? I have the hardware for it (128GB RAM) (Nvidia RTX 4060 TI 16GB VRAM).

#46
by CR2022 - opened

Can someone make quantized gguf or tell me how to do it? I have the hardware for it (128GB RAM) (Nvidia RTX 4060 TI 16GB VRAM).

Looks like someone has been making them here: https://civitai.com/models/1855105/rapid-wan-22-i2v-gguf

Can someone make quantized gguf or tell me how to do it? I have the hardware for it (128GB RAM) (Nvidia RTX 4060 TI 16GB VRAM).

City96 has instruction how to quant, maybe it helps you in future - https://github.com/city96/ComfyUI-GGUF/tree/main/tools Also there is something for ComfyUi quantizer: https://github.com/lum3on/ComfyUI-ModelQuantizer

Looks like someone has been making them here: https://civitai.com/models/1855105/rapid-wan-22-i2v-gguf

Thank you :)

Can someone make quantized gguf or tell me how to do it? I have the hardware for it (128GB RAM) (Nvidia RTX 4060 TI 16GB VRAM).

City96 has instruction how to quant, maybe it helps you in future - https://github.com/city96/ComfyUI-GGUF/tree/main/tools Also there is something for ComfyUi quantizer: https://github.com/lum3on/ComfyUI-ModelQuantizer

Thank you :)

CR2022 changed discussion status to closed

Sign up or log in to comment