Llama 3.2 1B was glacially slow at 0.0093 tok/sec ... which it hopes will be democratized thanks to BitNet. "BitNet is a transformer architecture that uses ternary weights," it explains.
Some results have been hidden because they may be inaccessible to you
Show inaccessible results