Focal Transformer: Focal Self-attention for Local-Global Interactions in Vision Transformers

❤️ Become The AI Epiphany Patreon ❤️ ► In this video I cover a new paper coming from Microsoft: “Focal Self-attention for Local-Global Interactions in Vision Transformers“ where they introduce a new transformer layer called focal attention. The main idea is to reduce the complexity but preserve the long-range dependencies. They achieve this by attending to the nearby tokens in a fine-grained manner and to the tokens that are further away they attend their coarsened representations. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ ✅ Paper: ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ ⌚️ Timetable:
Back to Top