Efficient Large-Scale Language Model Training on GPU Clusters Paper reading about Efficient Large-Scale Language Model Training on GPU Clusters. 2024-10-05 Distributed Training #Paper Reading
PipeFusion-Displaced Patch Pipeline Parallelism for Inference of DiT Models Paper reading of PipeFusion. 2024-09-29 Paper Reading #Diffusion Model Inference
xDiT Principle This is a brief introduction to the xDiT Principle. 2024-09-27 Distributed Training #Essay
Ring Attention Principle This is a brief introduction to the Ring Attention Principle. 2024-09-26 Distributed Training #Essay
Wafer-scale Computing Advancements, Challenges, and Future Perspectives Paper reading of wafer-scale chip. 2024-09-20 Paper Reading #Wafer-scale Chip
PMPP Learning-Chapter 15 Graph traversal Personal notebook 15 of Programming Massively Parallel Processors. 2024-09-18 CUDA #PMPP learning