At WWDC24, Apple unveiled Apple Intelligence, a new personal intelligence system integrated into iOS 18, iPadOS 18, and macOS Sequoia. Apple Intelligence is designed to empower users with intelligent tools for everyday tasks, leveraging advanced generative models that adapt on the fly to the user's current activity. The foundation models at the heart of Apple Intelligence are fine-tuned for diverse user experiences, including:
Apple Intelligence is built upon a foundation of groundbreaking privacy innovations and adheres to Apple's core values. The development process is guided by a set of Responsible AI principles:
Apple Intelligence relies on two primary foundation models:
These foundation models are part of a broader family of generative models developed by Apple, including:
Apple's foundation models are trained using the open-source AXLearn framework, leveraging JAX and XLA for efficiency and scalability. Training is conducted across various hardware platforms, including TPUs and GPUs, employing data parallelism, tensor parallelism, sequence parallelism, and FSDP to optimize the process. The training data consists of:
Apple emphasizes that user data is never used for training and applies filters to remove sensitive information from publicly available data.
Apple employs a hybrid data strategy incorporating human-annotated and synthetic data, along with rigorous data curation and filtering procedures. Two novel algorithms enhance the model's instruction-following abilities:
Apple has implemented various techniques to optimize its models for speed and efficiency, both on-device and in the private cloud. Key optimizations include:
These optimizations result in remarkable performance on iPhone 15 Pro, achieving a time-to-first-token latency of about 0.6 milliseconds per prompt token and a generation rate of 30 tokens per second.
Apple Intelligence leverages adapters, small neural network modules, to fine-tune its foundation models for specific tasks. These adapters dynamically specialize the models on-the-fly, adapting the attention matrices, attention projection matrix, and fully connected layers of the transformer architecture. This approach preserves the general knowledge of the base model while tailoring it to specific tasks. The adapter parameters are represented using 16 bits, with a typical size of 10s of megabytes for the on-device model.
Apple emphasizes human evaluation in benchmarking its models, as it closely correlates with user experience. Performance evaluations are conducted on both feature-specific adapters and foundation models. Apple's approach to performance evaluation is illustrated by the summarization adapter:
Apple's models with adapters demonstrate better summarization capabilities compared to a comparable model. However, the company acknowledges potential risks inherent in summarization, such as removing nuances or details. Extensive adversarial probing and continuous evaluation are conducted to mitigate potential harm.
Beyond feature-specific evaluations, Apple assesses the general capabilities of both on-device and server-based models using a comprehensive set of real-world prompts. The models are compared to open-source and commercial models of comparable size, demonstrating superior performance across various tasks, including brainstorming, classification, question answering, coding, and writing. Notably, Apple's on-device model outperforms larger models, while the server model compares favorably to commercial counterparts.
Apple also emphasizes safety and robustness by evaluating model performance on harmful content, sensitive topics, and factuality through adversarial prompts. The models consistently demonstrate lower violation rates compared to open-source and commercial models, highlighting their resilience against harmful inputs.
In addition to human evaluations, Apple utilizes benchmarks like Instruction-Following Eval (IFEval) to assess instruction-following capabilities. The results indicate that Apple's models excel at following detailed instructions compared to models of similar size.
Apple's foundation models and adapters underpin Apple Intelligence, a personal intelligence system deeply integrated into iPhone, iPad, and Mac. The system empowers users with advanced capabilities across language, images, actions, and personal context. Apple emphasizes responsible AI development, guided by its core values and a commitment to privacy. The company plans to share more information about its broader family of generative models in the future, expanding the possibilities of personal intelligence.
Ask anything...