Posts tagged #Links

Worth reading

A paper on pruning transformer attention heads during inference without retraining.