Versioned Late Materialization for Ultra-Long Sequence Training in Recommendation Systems at Scale

📰 ArXiv cs.AI

arXiv:2604.24806v1 Announce Type: cross Abstract: Modern Deep Learning Recommendation Models (DLRMs) follow scaling laws with sequence length, driving the frontier toward ultra-long User Interaction History (UIH). However, the industry-standard "Fat Row" paradigm, which pre-materializes these sequences into every training example, creates a storage and I/O wall where data infrastructure usage exceeds GPU training capacity due to data redundancy that is amplified in multi-tenant environments wher

Published 29 Apr 2026
Read full paper → ← Back to Reads