Skip Navigation

Niche Model of the Day: Openbuddy 25.2q, QwQ 32B with Quantization Aware Training

huggingface.co

OpenBuddy/openbuddy-qwq-32b-v25.2q-200k · Hugging Face

Seems there's not a lot of talk about relatively unknown finetunes these days, so I'll start posting more!

Openbuddy's been on my radar, but this one is very interesting: QwQ 32B, post-trained on openbuddy's dataset, apparently with QAT applied (though it's kinda unclear) and context-extended. Observations:

  • Quantized with exllamav2, it seems to show lower distortion levels than nomal QwQ. Its works conspicuously well at 4.0bpw and 3.5bpw.
  • Seems good at long context. Have not tested 200K, but it's quite excellent in the 64K range.
  • Works fine in English.
  • The chat template is funky. It seems to mix up the

    <think>

    and <|think|> tags in particular (why don't they just use ChatML?), and needs some wrangling with your own template.
  • Seems smart, can't say if it's better or worse than QwQ yet, other than it doesn't seem to "suffer" below 3.75bpw like QwQ does.

Also, I reposted this from /r/locallama, as I feel the community generally should going forward. With its spirit, it seems like we should be on Lemmy instead?

0 comments