Quests
#GPT
2 posts

History of LLM Evolution (5): Building the Path of Self-Attention — The Future of Language Models from Transformer to GPT
Building the Transformer architecture from scratch, deeply understanding core components like self-attention, multi-head attention, residual connections, and layer normalization.

The State of GPT
A structured overview of Andrej Karpathy's Microsoft Build 2023 talk, deeply understanding GPT's training process, development status, the current LLM ecosystem, and future outlook.