These are quantizations of the model ZwZ-8B, using a imatrix created from text_en_medium

Usage Notes:

  • Download the latest llama.cpp to use these quantizations.
  • Try to use the best quality you can run.
  • For the mmproj file, the F32 version is recommended for best results (F32 > BF16 > F16).
Downloads last month
1,409
GGUF
Model size
8B params
Architecture
qwen3vl
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for noctrex/ZwZ-8B-GGUF

Finetuned
inclusionAI/ZwZ-8B
Quantized
(4)
this model