March 6, 2020

[R] Fixed Encoder Self-Attention Patterns in Transformer-Based Machine Translation. They show that NMT models with manually engineered, fixed (i.e. position-based) attention patterns perform as well as models that learn how to attend.

submitted by /u/hardmaru [link] [comments] Source
March 6, 2020

[R] Batch Normalization Biases Deep Residual Networks Towards Shallow Paths

submitted by /u/koolaidman123 [link] [comments] Source
March 7, 2020

[R] “On Adaptive Attacks to Adversarial Example Defenses” – 13 published defenses at ICLR/ICML/NerIPS are broken

submitted by /u/Other-Top [link] [comments] Source
March 7, 2020

[D] Machine Learning – WAYR (What Are You Reading) – Week 82

This is a place to share machine learning research papers, journals, and articles that you're reading this week. If it relates to what you're researching, by […]