Attention Modules refer to modules that incorporate attention mechanisms. For example, multi-head attention is a module that incorporates multiple attention heads. Below you can find a continuously updating list of attention modules.
Method | Year | Papers |
---|---|---|
2017 | 18104 | |
2018 | 180 | |
2018 | 136 | |
2021 | 85 | |
2020 | 72 | |
2018 | 71 | |
2019 | 48 | |
2020 | 39 | |
2019 | 37 | |
2020 | 35 | |
2021 | 24 | |
2022 | 23 | |
2018 | 20 | |
2020 | 17 | |
2022 | 14 | |
2019 | 11 | |
2021 | 11 | |
2017 | 9 | |
2021 | 8 | |
2019 | 7 | |
2020 | 7 | |
2019 | 5 | |
2021 | 4 | |
2021 | 4 | |
2020 | 4 | |
2018 | 3 | |
2020 | 3 | |
2020 | 3 | |
2021 | 3 | |
2021 | 3 | |
2019 | 2 | |
2021 | 2 | |
2021 | 2 | |
2019 | 1 | |
2019 | 1 | |
2020 | 1 | |
2020 | 1 | |
2020 | 1 | |
2020 | 1 | |
2021 | 1 | |
2021 | 1 |