Jae Lee is the CEO and co-founder of Twelve Labs, a platform that gives businesses and developers access to multimodal video understanding.
The massive explosion of generative AI models for text and image has been unavoidable lately. As these models become increasingly capable, “foundation model” is a relatively new term being tossed around. So what is a foundation model?
The term remains somewhat vague. Some define it by the number of parameters, and therefore, how large a neural network is, and others by the number of unique and hard tasks that the model can perform. Is making AI models larger and larger and the model’s ability to tackle multiple tasks really that exciting? If you take away all the hype and marketing language, what is truly exciting about these new generations of AI models is this: They fundamentally changed the way we interface with computers and data. Think about companies like