Keeping a record about how I begin to pick up LLM
Doc. written in Obsidian
, which makes it inconvenient to make updates here due to multiple exclusive link format. I will update once and all when I finish and export html
or pdf
version.
- Architecture
- simple version & Basic idea
- overview
- Encoder, Decoder
- Attention
- Coding: Attention
- comprehensive version
- Encoder-only: BERT
- architecture
- pre-training
- Decoder-only: GPT
- Brief history: What's new, what's the differences
- pre-training
- tokenizer, embedding, transformer block, im_head
- Encoder-Decoder
- Comparation
- bert vs. GPT
- decoder-only vs. encoder-decoder
- Coding <- I'm here, have been preparing for an exam recently, catch up soon
- Encoder-only: BERT
- simple version & Basic idea
- Training process
- Overview
- [ every step ]
- Use LLM
- 实战1
- 实战2
- Research