Self-Improving Pretraining: using post-trained models to pretrain better models
By: Ellen Xiaoqing Tan, Shehzaad Dhuliawala, Jing Xu
Published: 2026-01-29
View on arXiv →#cs.AI
Abstract
The "Self-Improving Pretraining" framework integrates alignment objectives (safety, factuality, quality) directly into LLM pretraining using a powerful post-trained model as a dynamic rewriter and judge. This method leads to significant gains in generation coherence and factuality, improving the reliability and trustworthiness of large language models for real-world use.