Hacker News new | past | comments | ask | show | jobs | submit login
Nyströmformer: A Nyström-Based Algorithm for Approximating Self-Attention (arxiv.org)
85 points by tmfi on Feb 11, 2021 | hide | past | favorite | 2 comments



Here's a nice video by Yannick Kilcher explaning the Nystromformer: https://www.youtube.com/watch?v=m-zrcmRd7E4

The benefits over regular transformers is that it is more efficient (does less operations), as the original transformer has a quadratic complexity in the number of input tokens.


It also links to a comparison that is not in the paper, against Performer, Linformer and Reformer:

https://twitter.com/tanmingxing/status/1359301186734620675




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: