I can’t believe after my gripe yesterday that I have picked today’s paper. This paper claims that a feed-forward network with self-attention trains faster and performs better than a recurrent or a convolutional neural network on translation tasks. I have to come back to this at a later time though because I need to brush up on much of the terminology.
-
Recent Posts
Recent Comments
Archives
Categories
Meta
Pingback: Paper: Concepts in a Probabilistic Language of Thought (64/365) | Latent observations