Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yeah, in some sense any layer is a "memory module". Perhaps more specifically, attention solves the problem of directly correlating two items in a sequence that are very, very far away from each other. I'd generally caution against using attention prematurely as it's extremely slow, meaning you'll waste a lot of your time and resources without knowing if it'll help. Stacking conv layers or using recurrence is an easy middle step that, if it helps, can guide you on whether attention could provide even more gains.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: