The Rapid Growth of Hugging Face Models: Trends and Insights
We are seeing almost 600,000 unique users downloading models from Hugging Face, which is a significant increase in recent times. This trend is quite notable, and it's essential to understand what's driving this growth. One plot that showcases this trend is the number of unique users downloading models. In the pre-plot, we can see the number of unique users unloading the model, while the current plot shows the number of Llama-based repositories being created on Hugging Face. The former indicates the popularity of certain models, whereas the latter highlights the proliferation of new Llama-based models.
Llama is an architecture that has gained immense popularity in recent times. However, it's essential to note that not all Llama-based models are official ones. Many people have been training and sharing fine-tuned models with similar architectures but without being officially affiliated with the Llama framework. This proliferation of new models is indeed impressive, with between 3,000 to 4,000 new Llama-based models being created every week. Moreover, the plot shows a Lama-shaped trend, indicating a growth curve that's not unlike a typical startup trend.
Another significant trend we're witnessing is the emergence of small but powerful models. Just a few years ago, there was a limited set of very small models available. However, now we see a plethora of powerful models at small scales. For instance, there are GMA-5 models in both text and vision spaces, as well as various smaller variants. Additionally, models like Whisper, PEAR, and MobileBERT have made significant strides in speech recognition and computer vision tasks. Furthermore, there are 20 different variations of MobileBERT alone, showcasing the vast range of smaller models that can be achieved while maintaining high performance.
While Hugging Face is renowned for its Transformers-based models, it's essential to note that the platform hosts a wide range of models from various libraries. These include reinforcement learning, computer vision, and more. The model download plot showcases this diversity, with a significant number of users opting for models from different libraries like FAIR (facebook ai research), SQuAD, and more. Furthermore, Hugging Face has an impressive collection of over 1.5 million models in total.
The creation of new models is another trend that's worth highlighting. According to the plot, every single week, between 30,000 and 40,000 new models are being shared on Hugging Face. This is a significant increase compared to when our founder joined Hugging Face about three-and-a-half years ago, when there were only a few thousand new models being created weekly. This growth in model creation is indeed remarkable.
The Architecture Life Cycle: Understanding the Rise and Fall of Models
When a new model architecture emerges, it often sparks excitement among researchers and practitioners alike. However, this excitement can be short-lived if the model doesn't gain widespread adoption or longevity. There are typically two possible outcomes for a new model's life cycle: it either gains long-term usage or adoption or fails to do so.
To help extend a model's life cycle and foster its adoption, there are several strategies that can be employed. Firstly, focusing on a single architecture rather than constantly changing or experimenting with different architectures can help maintain user interest. Llama is an excellent example of such an architecture, which has become synonymous with the rapid growth of generative text models.
Secondly, transparency and openness about the model's training data and usage are crucial for its adoption. Not all newly created models share their training data or provide clear information on how to use them. This lack of transparency can lead to a decrease in user interest and adoption over time.
Lastly, avoiding complex licenses that might deter users from adopting the model is essential. A simple and consistent API that aligns with other Hugging Face models can go a long way in making it easier for users to integrate the new model into their workflow. Avoiding complicated licensing agreements will also make it more likely for new models to be adopted, which is crucial for any open-source platform like Hugging Face.
In conclusion, understanding the trends and insights behind the rapid growth of Hugging Face models can provide valuable lessons for researchers and practitioners alike. By focusing on a single architecture, maintaining transparency, avoiding complex licenses, and promoting consistency, we can foster a more inclusive community around model creation and adoption.