When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models.

ICML(2024)

引用 0|浏览31
暂无评分
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要