News

MM1 is a multimodal large language model, or MLLM, meaning it is trained on images as well as text. This allows the model to respond to text prompts and also answer complex questions about ...
Released as a preprint research paper, MM1 essentially offers a new method for using AI-generated data and labels to speed up training of new models — including possibly Siri 2.0. What is Apple MM1?
Apple researchers say they’ve developed a family of multimodal models — which refers to an AI system that can interpret and generate different types of data, such as text and images at the ...
Apple says it plans to use a diverse dataset that includes interleaved image-text documents, image-caption pairs, and text-only data to help train and develop MM1. This, Apple claims, should allow ...
Being multimodal, MM1 is capable of working with both text and images. Overall, its capabilities and design are similar to the likes of Google’s Gemini or Meta’s open-source LLM Llama 2.
Apple released MM1, and they have released more information than most open source LLM companies. I have been testing MLX for some of my workflows, it’s probably the fastest among other python ...
Apple has revealed its latest development in artificial intelligence (AI) large language model (LLM), introducing the MM1 family of multimodal models capable of interpreting both images and text data.