Processing and understanding human language with computational methods.
Breaking text into tokens — the bridge between raw text and model input.
The architecture behind modern language models — attention, encoders, and decoders.
Reinforcement Learning
Tokenization