I read a lot about foundation model and large language model.
However, I dont find a clear definition what exactly is a foundation model. Is large language model and foundation model the same thing?
I read a lot about foundation model and large language model.
However, I dont find a clear definition what exactly is a foundation model. Is large language model and foundation model the same thing?
At this point in time, there does not appear to be a really widely-agreed-upon definition of "Foundation models". If you want one, the best place to go would be this paper from Stanford that coined the term. Generally at least some of the following ideas apply (according to some people they must all apply, according to others only some need to apply):
Large Language Models would typically be trained specifically on language-related data (text). So, I suppose an LLM could sometimes serve as a Foundation model, but it's not necessarily the same thing.
NOTE: I would like to remark that, from many experts outside of Stanford, there is quite a bit of pushback against the particular term "Foundation model". There seems to be a bit of PR involved, where it seems like they may like to coin a new term for something that really doesn't need a new term. Just saying it's, for example, a large pre-trained model, may be more clear.