Non-Parametric Transformers - Paper Explained

Non-Parametric Transformers - Paper Explained

Aleksa Gordić - The AI Epiphany via YouTube Direct link

Key ideas of the paper

1 of 15

1 of 15

Key ideas of the paper

Class Central Classrooms beta

YouTube playlists curated by Class Central.

Classroom Contents

Non-Parametric Transformers - Paper Explained

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Key ideas of the paper
  2. 2 Abstract
  3. 3 Note on k-NN non-parametric machine learning
  4. 4 Data and NPT setup explained
  5. 5 NPT loss is inspired by BERT
  6. 6 A high-level architecture overview
  7. 7 NPT jointly learns imputation and prediction
  8. 8 Architecture deep dive input embeddings, etc
  9. 9 More details on the stochastic masking loss
  10. 10 Connections to Graph Neural Networks and CNNs
  11. 11 NPT achieves great results on tabular data benchmarks
  12. 12 NPT learns the underlying relational, causal mechanisms
  13. 13 NPT does rely on other datapoints
  14. 14 NPT attends to similar vectors
  15. 15 Conclusions

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.