Another issue with continual learning is that it likely doesn’t have the efficiency of today’s cloud-based LLMs:
inference-time gradient computation and per-user weight divergence breaking the efficiency of batched serving. (Are We in a Continual Learning Overhang?)
Another issue with continual learning is that it likely doesn’t have the efficiency of today’s cloud-based LLMs: