[1-Bit Quantization] Scaling Binary (BitNet) Logic for 16MB Parameter Golf Constraint #833
Zzcvbnma
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi Team,
I'm Ranjit, an independent AI developer. I'm taking on the 16-million parameter (16MB) challenge by implementing a 1-bit Binary (BitNet-inspired) architecture.
To strictly adhere to the model size constraint while maximizing parameter count, I've designed a custom
BitLinearlayer that uses +1/-1 weights. This approach allows me to squeeze ~134M parameters within the 16MB artifact limit.Current Progress:
Zzcvbnma/Parameter-Golf-Binary-BeastI'm aiming for sub-1.10 BPB loss. Has anyone else experimented with extreme quantization for this challenge? I'd love to hear your thoughts on convergence stability in the first 600 seconds!
Beta Was this translation helpful? Give feedback.
All reactions