Non-Parametric Transformers | Paper explained

In this video I do a deep dive of the “Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning“ paper. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ ✅ Non-Parametric Transformers paper: ✅ Jay Alammar’s BERT blog: ✅ My LinkedIn post (Judea Pearl): (also check out my other posts I made related to this) ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ ⌚️ Timetable: 00:00 Key ideas of the paper 01:40 Abstract 02:55 Note on k-NN (non-parametric machine learning) 04:30
Back to Top