Apple is not so secretive? OpenELM and DCLM
Apple, the tech giant notorious for its secretive approach to product development, has made a surprising move in the AI landscape with the release of OpenELM, their state-of-the-art open language model. This unexpected move challenges the company's usual closed-off nature, showcasing a commitment to transparency and collaboration in the AI community.
OpenELM: A Paradigm Shift in Apple's AI Strategy
Contrary to popular belief, OpenELM embraces the principles of reproducibility and transparency, which often need to be improved in large language models. This open-source model aims to advance AI research, ensure the trustworthiness of results, and facilitate investigations into potential data and model biases.
Impressive Performance and Efficiency
OpenELM's performance is nothing short of remarkable. It demonstrated a 2.36% increase in accuracy compared to the OLMo model while requiring only half the pre-training tokens. Apple's brilliant team of researchers achieved this feat through a technique called layerwise scaling, which optimizes parameter allocation across the model's architecture. This innovative approach enhances data processing efficiency and improves accuracy, setting OpenELM apart from other models.
Trained on a Wealth of Public Data
OpenELM's foundation lies in a massive dataset from public resources like GitHub, Wikipedia, Stack Exchange, and others, totaling billions of data points. This extensive training data contributes to the model's impressive capabilities. Moreover, Apple provides all the necessary tools and frameworks for further training, empowering researchers and developers to customize and enhance the model.
A Comprehensive Open-Source Framework
Apple has made OpenELM a comprehensive open-source framework, encompassing not only model weights and code for execution but also training logs, checkpoints, and pre-training configurations. This level of openness goes beyond typical practices, encouraging wider participation in AI research and development.
The Road Ahead: Challenges and Opportunities
While OpenELM boasts impressive accuracy, it currently faces challenges in terms of speed due to the complex algorithms involved. However, Apple is actively working on improving this aspect, and aspiring hackers could even contribute to this effort as an exciting internship or academic project.
Apple's Bold Step Towards Open Innovation
Based on this, Apple has entered the open-source AI arena with a bang, releasing the DCLM 7B. Based on OpenELM and trained on 2.5 trillion tokens, this model achieves a remarkable 63.72 on the MMLU benchmark, surpassing Mistral’s 7B. It’s a testament to Apple’s commitment to advancing open-source machine learning. OpenELM marks a significant shift in Apple's approach to AI, aligning with the company's core philosophy of "Making things revolutionary and better than anyone can think of." By embracing open-source principles and fostering collaboration, Apple is poised to drive innovation and push the boundaries of AI research and development.
Recommended by LinkedIn
When they are so progressing on this front, why OpenAI? In the future, will they do it like they did previously with so many other partnerships? What is your opinion? Share your thoughts
For those who are interested to play and know more about these models, here are the details:
OpenELM
DCLM
I help mid-market companies and start-ups create and align their product & corporate strategies | $2B+ Products Launched | Innovation Leader | Design-Thinker | Builder
4moRama, thanks for sharing.