If the ‘Apple Glow’ Event Were an Email...
While some are busy calling the iPhone 16 a mere upgrade, the real insanity lies in the A18 Bionic chip. This powerhouse not only boosts performance with its next-gen neural engine but also fuels advanced AI features, all while keeping your data safe with federated learning. And it’s built on a 3nm process, making it ridiculously efficient and secure.
But wait, it gets even crazier! Apple’s Intelligence Adapters for priority notifications and summaries are now fully inspectable, giving users more control than ever before.
The devil’s in the details
Citing fellow ML researcher at Hugging Face, Cyril Zakka, who brought this to light, Philipp Schimid, the technical lead and LLMs at Hugging Face, said that the available 3B on-device Apple Intelligence model has special prompts and adapters for features like priority notifications, localisation, guardrails, etc.
For example:
Apple Intelligence Coming Next Month
Apple announced that it will be bringing Apple Intelligence next month with iOS 18.1, iPadOS 18.1, and macOS Sequoia 15.1, offering advanced writing tools, visual intelligence (aka Google Lense clone), a smarter Siri, enhanced photo features, and on-device privacy.
In June, when Apple officially announced the launch of ‘Apple Intelligence’, the former member of OpenAI, Andrej Karpathy, too lauded its attempt to prioritise privacy in its integration of AI into its OS and called out seven major themes observed: multimodal I/O, agentic, frictionless, initiative, delegation hierarchy, modularity, and, lastly, privacy.
“We’re quickly heading into a world where you can open up your phone and just say stuff. It talks back, and it knows you. And it just works. Super exciting, and as a user, quite looking forward to it,” said Karpathy.
Private Cloud Compute extends the privacy and security of Apple devices into the cloud to unlock even more intelligence.
Recommended by LinkedIn
For the most part, Apple’s on-device processing will take care of your needs. However, there may be times when additional data is required to answer your query, and this is where ‘Private Cloud Compute’ comes into the picture.
With Private Cloud Compute, Apple Intelligence can scale its computational capacity and utilise larger, server-based models for more complex requests. These models run on Apple silicon-powered servers, ensuring that data is never retained or exposed.
Apple says that when a user makes a request, Apple Intelligence analyses whether it can be processed on-device. It can draw on ‘Private Cloud Compute’ and send only the relevant data if greater computational capacity is needed.
Apple has enabled cryptography encryption, so your Apple device will never be able to talk to the Apple server unless the software is publicly logged for inspection.
Meanwhile, Elon Musk has raised concerns over the OpenAI-Apple partnership. “If Apple integrates OpenAI at the OS level, then Apple devices will be banned at my companies. That is an unacceptable security violation,” posted Musk on X.
Now You Can Train LLMs on a Two-Year-Old Desktop-Grade NVIDIA 3090 GPU
Recently, researchers from the School of Computer Science & Technology, Soochow University, released a research paper titled ‘MemLong: Memory-Augmented Retrieval for Long Text Modeling’, where they successfully extended an LLM’s context window from 2k to 80k tokens on a two-year-old desktop-grade NVIDIA 3090 GPU.
This opens new horizons for users with limited hardware access and who still want to use AI applications locally on their computers. Read on.
AI Bytes