Self-Improving Pretraining: using post-trained models to pretrain better models

By: Ellen Xiaoqing Tan, Shehzaad Dhuliawala, Jing Xu

Published: 2026-01-29

View on arXiv →
#cs.AI

Abstract

The "Self-Improving Pretraining" framework integrates alignment objectives (safety, factuality, quality) directly into LLM pretraining using a powerful post-trained model as a dynamic rewriter and judge. This method leads to significant gains in generation coherence and factuality, improving the reliability and trustworthiness of large language models for real-world use.

FEEDBACK

Projects

No projects yet