tgoop.com/CodeProgrammer/4257
Last Update:
🎓 Stanford has released a new course: “Transformers & Large Language Models”
The authors are the Amidi brothers, and three free lectures are already available on YouTube. This is probably one of the most systematic introductory courses on modern LLMs.
Course content:
• Transformers: tokenization, embeddings, attention, architecture
• #LLM basics: Mixture of Experts, decoding types
• Training and fine-tuning: SFT, RL, LoRA
• Model evaluation: LLM/VLM-as-a-judge, best practices
• Tricks: RoPE, attention approximations, quantization
• Reasoning: scaling during training and inference
• Agentic approaches: #RAG, tool calling
If you are already familiar with this topic — it’s a great opportunity to refresh your knowledge and try implementing some techniques from scratch.
https://cme295.stanford.edu/syllabus/
https://www.tgoop.com/CodeProgrammer