continuity is a function of memory. although model distillation uses the term knowledge, it’s the same concept. it might not apply to current models, but i suspect at some point future models will essentially be ‘training’ 24⁄7, the way the human mind uses new experiences to update it’s neural connections instead of simply updating working memory.
continuity is a function of memory. although model distillation uses the term knowledge, it’s the same concept. it might not apply to current models, but i suspect at some point future models will essentially be ‘training’ 24⁄7, the way the human mind uses new experiences to update it’s neural connections instead of simply updating working memory.