LLM Learning Initial

LLM Learning Initial

Welcome to the world of Large Language Models

For this summer vacation, I will begin the learning process for the fundamental principle and downstream techniques and applications for LLM (Large Language Models). I hope to find the journey interesting and fruitful!

All the code will be refactored and categorized into LLM Learn

LLM Learning Materials

LLM Learning Contents

  • Basic architecture for Large Language Models

    • Attention Mechanism (Attention is all you need!) ✅

    • RNN, LSTM, GRU (will be covered in the future)

    • Seq2Seq Model ✅

    • Transformer Architecture ✅

    • Other basic NLP knowledge (word embeddings, etc.)

  • Pre Training for LLM

    • Loading Datasets

    • Self-supervised Learning

    • More advanced architecture for LLM pre-training, see advanced structure part.

  • Post Training for LLM

    • Quantization for Model Optimization

    • Knowledge Distillation

    • Fine-tuning Techniques

      • SFT
      • RFT
      • RLHF (Reinforcement Learning from Human Feedback)
    • LLM Evaluation

  • Advanced Structure for LLM

  • Test time compute for LLM (after training)

  • LLM DownStream Applications

    • This section will be recorded in the future.

    • RAG

    • LangChain Community

Updating Status

  • 2025/07/28: Finish two long-standing blog posts: AINN-Attention & AINN-Transformer

    • Finish tutorial for basic Attention mechanism and Transformer Structure.

Current Todo List

  • Finish the implementation code of Transformer Module in dl2ai.

  • Learning courses: Word Embedding and basic NLP knowledge.


LLM Learning Initial
https://xiyuanyang-code.github.io/posts/LLM-Learning-Initial/
Author
Xiyuan Yang
Posted on
July 27, 2025
Updated on
August 2, 2025
Licensed under