Text Generation
Transformers
Safetensors
qwen3_moe
conversational

Can somebody please make GGUF or MLX version quickly?

#1
by Narutoouz - opened

Chinese AI models are pushing the AI Industry forward. Though I must say it was done due to the data obtained from GPT OSS 120b model from OpenAI. As an AI student, more over a consumer of AI, thanks for democratizing this by open sourcing this, Alibaba-Aspara. Again, thanks to OpenAI for Opensourcing GPT OSS 120b, it is still an excellent model, even though it been over 6 months since it was released. Sam Altman's promise to open source was real and thanks for that.

The benchamarks are already SOTA in the preview version compared to other models in this parameter range, I can't wait to see the instruct and thinking variants of this model. Waiting excitely for those!

Narutoouz:

  • you can easily produce a GGUF version, without even needing to download the model and run the command by yourself, using https://huggingface.co/spaces/ggml-org/gguf-my-repo as Qwen3MoeForCausalLM (you can find the architecture in config.json) has already been supported for some times now
  • I don't know how you missed it, but this is indeed a thinking model, as per the README and the name of the model!
  • beware that this is currently a text only model, it hasn't been trained for tool usage, so even if the base model was I guess you shouldn't expect anything

Chinese AI models were pushing the frontier long before GPT OSS was released.

Sign up or log in to comment