Term: Checkpointing
Day: 10
Level: Fluency
Category: Training & Optimization
🪄 Simple Definition:
Saving an AI model’s progress during training so it can be resumed or reused later without starting over.
🌟 Expanded Definition:
Checkpointing stores the state of a model—its learned weights, parameters, and progress—at certain points during training. This prevents wasted time if training is interrupted and allows developers to reuse or fine-tune the model from a saved point. It’s especially valuable for large models that take weeks or months to train.
⚡ In Action:
A research team training a large language model saves a checkpoint every 24 hours. If their servers crash, they can restart from the last checkpoint instead of losing weeks of progress.
đź’ˇ AIS+ Pro Tip:
Use checkpointing strategically — frequent saves add storage overhead, but infrequent saves risk lost progress. For large-scale projects, pair checkpointing with distributed training to balance efficiency and reliability.
🔍To find all posted terms, simply search for the phrase “Daily Dose” in the AIS+ community.
Start AI Terms Daily Dose from Beginning:
📣 Complementary Series 📣
📚 AI Terms Everyone Should Know Series
Your complete guide to mastering AI vocabulary, from basics to advanced, with context and real-world examples by @Michael Wacht Series inspired by the "Only Cheatsheet to Master AI Basics" post by @Yash Chauhan found at: