NRI (Nomura Research Institute)’s Post

The current wave of generative #AI models is built on the Transformer architecture, which has been popularized by the emergence of large language models (LLMs). Despite their prominence, LLMs have inherent drawbacks and constraints. To address these issues, researchers are now focusing on developing smaller language models that could potentially revolutionize the field of generative AI. Read more. #GenAI #GenerativeAI #LanguageModels

I think there is an interesting conversation to be had around parameters vs tokens and delivering smaller models that can operate on the edge. Companies have been more focused on large parameter models, but there is interesting work happening on lower parameter models with more training to produce models that compete with higher parameter in a smaller package.

To view or add a comment, sign in

Explore topics