Pretraining 중심의 LLM Training Stack 표준화 및 LLM-OS 아키텍처 정립
The Man Who Summoned Ghosts | Chapter 2: The Training Stack Is Not a Secret
The Man Who Summoned Ghosts | Chapter 2: The Training Stack Is Not a Secret
EMO: Pretraining mixture of experts for emergent modularity
Open-source AI I'm watching: DeepSeek V4, VibeVoice, and the n8n effect
Even 'uncensored' models can't say what they want
Cosmopedia: how to create large-scale synthetic data for pre-training Large Language Models