Data Mixing Agent: Learning to Re-weight Domains for Continual Pre-training

📰 ArXiv cs.AI

arXiv:2507.15640v2 Announce Type: replace-cross Abstract: Continual pre-training on small-scale task-specific data is an effective method for improving large language models in new target fields, yet it risks catastrophic forgetting of their original capabilities. A common solution is to re-weight training data mixtures from source and target fields on a domain space to achieve balanced performance. Previous domain reweighting strategies rely on manual designation with certain heuristics based o

Published 14 Apr 2026
Read full paper → ← Back to Reads