AXIOM uses a novel weight representation combining lossless compression with a built-in optimizer. Weights stored at 0.62 bytes/param vs 4 bytes FP32. Gradient updates happen directly in compressed space.
Not quantization-aware training or LoRA — every parameter fully trainable, convergence matches AdamW.
Solo founder from Canada. Self-taught CUDA/ML. Applying to YC S26.
Happy to answer questions.
Put two spaces at the beginning of a line for monospace.
Like this
Thanks for the tip! It won't let me edit it. I think I will hide and repost it does look sloppy.
I built QuarterBit because AI training costs are insane. A 70B model needs 840GB of memory — that's 11 A100 GPUs at $30+/hour.
QuarterBit AXIOM compresses training memory 15x. Same model. Same quality. Fraction of the hardware.
RESULTS:
91% energy reduction vs standard training. 100% trainable weights (not LoRA/adapters). 3 lines of code.HOW IT WORKS:
TRY IT: Demo (FREE): https://www.kaggle.com/code/kyleclouthier/quarterbit-axiom-1...Benchmarks: https://quarterbit.dev
AXIOM uses a novel weight representation combining lossless compression with a built-in optimizer. Weights stored at 0.62 bytes/param vs 4 bytes FP32. Gradient updates happen directly in compressed space.
Not quantization-aware training or LoRA — every parameter fully trainable, convergence matches AdamW.
Solo founder from Canada. Self-taught CUDA/ML. Applying to YC S26.
Happy to answer questions.
Put two spaces at the beginning of a line for monospace.
Thanks for the tip! It won't let me edit it. I think I will hide and repost it does look sloppy.