Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
auggierose
3 months ago
|
parent
|
context
|
favorite
| on:
Writing an LLM from scratch, part 13 – attention h...
No, it doesn't really clarify things. I had the best linear algebra grades in my year at my university, and if you don't know anything about kernels, this is not helpful (what are xi and yi in the first place?).
lambdasquirrel
3 months ago
[–]
It's all described in the referenced link. No need for everyone to get antsy.
> 0.
http://bactra.org/notebooks/nn-attention-and-transformers.ht
...
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: