by kleene_op on 12/26/2023, 2:48:40 AM with 5 comments
Excuse me for this potentially dumb question, but..
Why don't we train LLMs with user inputs at each step instead of keeping the model static and feeding the whole damn history everytime?
I think I may have a clue as to why actually: Is it because this would force us to duplicate the model for every user (since their weights would diverge) and company like OpenAI deem it too costly?
If so, will the rise in affordability of local models downloaded by individuals enable the switch for the continuous training approach soon enough or am I forgetting something?
Excuse me for this potentially dumb question, but..
Why don't we train LLMs with user inputs at each step instead of keeping the model static and feeding the whole damn history everytime?
I think I may have a clue as to why actually: Is it because this would force us to duplicate the model for every user (since their weights would diverge) and company like OpenAI deem it too costly?
If so, will the rise in affordability of local models downloaded by individuals enable the switch for the continuous training approach soon enough or am I forgetting something?