mirror of
https://github.com/rasbt/LLMs-from-scratch.git
synced 2024-11-25 16:22:50 +08:00
1183fd7837
Some checks failed
Code tests (Linux) / test (push) Has been cancelled
Code tests (macOS) / test (push) Has been cancelled
Test PyTorch 2.0 and 2.5 / test (2.0.1) (push) Has been cancelled
Test PyTorch 2.0 and 2.5 / test (2.5.0) (push) Has been cancelled
Code tests (Windows) / test (push) Has been cancelled
Check hyperlinks / test (push) Has been cancelled
Spell Check / spellcheck (push) Has been cancelled
PEP8 Style checks / flake8 (push) Has been cancelled
|
||
---|---|---|
.. | ||
01_main-chapter-code | ||
02_bonus_efficient-multihead-attention | ||
03_understanding-buffers | ||
README.md |
Chapter 3: Coding Attention Mechanisms
Main Chapter Code
- 01_main-chapter-code contains the main chapter code.
Bonus Materials
- 02_bonus_efficient-multihead-attention implements and compares different implementation variants of multihead-attention
- 03_understanding-buffers explains the idea behind PyTorch buffers, which are used to implement the causal attention mechanism in chapter 3