Sparse Llama: 70% Smaller, 3x Faster, Full Accuracy

  • Specifically this is Llama2, not Llama3, was a bit disappointed from that. Also wasn't totally clear from the article - will this actually increase GPU inference speed / decrease GPU memory usage?