Apple Unveils OpenELM: AI Language Models for Smartphones

Apple Unveils OpenELM: AI Language Models for Smartphones

By
Maksim Ivanov
2 min read

Apple introduced a new set of small AI language models called OpenELM, which can be run directly on smartphones. These models are currently available on Hugging Face under the Apple Sample Code License. The models are mostly proof-of-concept research models and come in two flavors: pretrained and instruction-tuned. Apple aims to empower the open research community by releasing the source code, model weights, and training materials. While the company has not yet integrated these models into its consumer devices, the upcoming iOS 18 update is rumored to include new AI features that utilize on-device processing to ensure user privacy.

Key Takeaways

  • Apple introduces OpenELM, small source-available AI language models for local device use.
  • The OpenELM models range from 270 million to 3 billion parameters and come as pretrained or instruction-tuned.
  • Apple's OpenELM features a 2048-token maximum context window and was trained on around 1.8 trillion tokens of data.
  • The models utilize a "layer-wise scaling strategy" and reportedly achieve a 2.36 percent improvement in accuracy over a similar model.
  • The upcoming iOS 18 update is rumored to include new AI features utilizing on-device processing to ensure user privacy.

Analysis

Apple's introduction of OpenELM, small AI language models for local device use, showcases a strategic move to empower the open research community with proof-of-concept models. This initiative reflects Apple's commitment to privacy and on-device processing, possibly impacting AI research organizations, app developers, and the broader AI community. The release of source code and model weights can catalyze advancements in AI research, while the rumored AI features in the upcoming iOS 18 update could potentially enhance user privacy and elevate Apple's competitive edge in the smartphone market. This development may lead to a shift towards on-device AI processing, influencing the landscape of AI application development and privacy protection in the long-term.

Did You Know?

  • OpenELM:

    • Small source-available AI language models for local device use introduced by Apple.
    • Ranging from 270 million to 3 billion parameters and available as pretrained or instruction-tuned.
    • Empowers the open research community by releasing source code, model weights, and training materials.
  • 2048-token Maximum Context Window:

    • Feature of Apple's OpenELM models allowing the consideration of a maximum context of 2048 tokens during language processing.
    • Trained on around 1.8 trillion tokens of data, indicating its capability to handle large volumes of information.
  • Layer-wise scaling strategy:

    • Utilized by the OpenELM models to achieve reported improvements in accuracy over similar models.
    • Signifies a specialized approach to scaling neural network layers for enhanced performance.

You May Also Like

This article is submitted by our user under the News Submission Rules and Guidelines. The cover photo is computer generated art for illustrative purposes only; not indicative of factual content. If you believe this article infringes upon copyright rights, please do not hesitate to report it by sending an email to us. Your vigilance and cooperation are invaluable in helping us maintain a respectful and legally compliant community.

Subscribe to our Newsletter

Get the latest in enterprise business and tech with exclusive peeks at our new offerings