Drastically larger, with performance to match.
Upgraded Jinja template too.
DavidAU/Qwen3.5-40B-Claude-4.5-Opus-High-Reasoning-Thinking
UPDATE:
All of these are now up; and can be downloaded.
Awaiting quants.
RE: 13B:
=> one is upscaled + trained, the other is merge of two 9Bs fine tunes (and upscaled).
They are hidden as of this writing (undergoing private testing), awaiting final metrics / eval.
If they "pass" ; they will be made public.
These will be active within 24-48 hrs pending results.
Currently have full running 13B (GLM 4.7 Flash) - which is very strong ; and experimental 21Bs of Qwen 3.5.
These are trained.
These are in testing, and access is limited as of this writing.
As for MOEs:
This is a little more complicated as scripting must be written for Mergekit to "moe together" 0.8B, 2B, 4B, 9Bs etc etc.
A draft (by me) has been completed to do this; but not tested/debugged yet.
No time line here ; too many variables.
RE 35B moes ; it is possible to address this in a different way ; but I have not tried it yet.
This is a different approach than REAP.
9 Heretic Uncensored LFM fine tunes are now up at my repo:
https://huggingface.co/DavidAU/models?sort=created&search=lfm
Model card updates in progress as I write this.
The merges will take a wee bit longer.
...and 5 more new "non-heretic" ones too.
@muxodious ; Excellent.
In the cue.
Important note:
I can make the base models W reasoning datasets; however the "Kimi Mega Brain" is a complex merge of these base models (trained with different datasets) by Nightmedia.
I will query Nightmedia to see if he will do an updated "Heretic" mega brain merge after the "heretic" versions are complete.
Waiting for updates W HEretic/Transformers to make this possible with "thinking" LFM base.
For each model ; the quants are listed under quantizations.
Hey;
I am currently restricting access to the source presently due to past issues with abuse of the source (of my models), which lead to community issues due to non-disclosure of tech details of the model as well as issues related to non-attribution of multiple parties.
I may release it in a few weeks.
The 80Bs will soon be on the docket.
Issue with ablits -; may be some losses in brain dept so to speak.
Ablits are much better than they used to be, but when it comes to tuning - can be a bit of a nightmare.
Received your message, you can contact me via discord:
David_AU [note underscore]
Or open/setup a model on your Hugging face and I will contact you via community tab there if you prefer.
Hey;
I am DavidAU
[ https://huggingface.co/DavidAU ]
Also in the fine Perth, WA area ; please take a look at my repo and see if there is something we have in common. (?)
Your desc of "Algorithm in progress" is not too clear.
I build models including quants, merges and fine tunes.
You can contact me on DISCORD too:
David_AU
Cheers.
David