Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
2
3
martin ohnout
ohni
Follow
shayzinasimulation's profile picture
1 follower
·
4 following
ohni
AI & ML interests
None yet
Recent Activity
reacted
to
DavidAU
's
post
with ❤️
3 days ago
SAVANT COMMANDER: 48B-A4B , 256k Context, GATED MOE. I am going to showcase some other people's tuning work, that I have put into a GATED Distill MOE (Qwen3) ; 256 K context. Special thanks to all the tuners (listed in the model tree and repo page with special shoutout to "TeichAI" - using Unsloth for a lot of the Distills in this model): Savant Commander is a specialized MOE model that allows you to control which expert(s) (of 12) are assigned to your use case(s) / prompt(s) ... directly (by name(s)), as opposed to having the "choices" made for you. The model is composed of 12 DISTILLS (compressed 12x4B MOE) of top closed (GPT5.1, OpenAI 120 GPT Oss, Gemini (3), Claude (2) ) and open source models (Kimi, GLM, Deepseek, Command-A, JanV1 ) all in one. 256k Context, 2 experts activated. PS: There is also a "heretic" / "decensored" version too ; listed on this model page. https://huggingface.co/DavidAU/Qwen3-48B-A4B-Savant-Commander-GATED-12x-Closed-Open-Source-Distill-GGUF
upvoted
a
collection
8 days ago
Qwen 3 / 2.5 Reasoning/Thinking REG + MOEs.
upvoted
a
collection
2 months ago
Qwen3 - 30B-A3B (128 experts) and higher
View all activity
Organizations
None yet
ohni
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a model
8 months ago
google/medgemma-4b-it
Image-Text-to-Text
•
4B
•
Updated
Oct 28, 2025
•
423k
•
860
liked
a dataset
12 months ago
open-thoughts/OpenThoughts-114k
Viewer
•
Updated
Aug 31, 2025
•
228k
•
71.7k
•
790
liked
a model
12 months ago
HKUSTAudio/Llasa-3B
Text-to-Speech
•
4B
•
Updated
May 10, 2025
•
979
•
524