--- frameworks: - pytorch license: apache-2.0 tasks: - text-generation language: - en base_model: - Llama/Llama-3.1-8B-Instruct pipeline_tag: text-generation library_name: transformers --- [Elastic Attention: Test-time Adaptive Sparsity Ratios for Efficient Transformers](https://www.arxiv.org/abs/2601.17367)