论文地址:
https://arxiv.org/abs/2108.02347arrow-up-right
利用Local Attention + Low-rank Attention逼近Softmax Attention,其中Low-rank Attention就是常用的Linear Attention。
https://github.com/minhtannguyen/fmmformer-code-submissionarrow-up-right
简单常规的思路,类似的论文也不少了。
Last updated 2 years ago