Microsoft has released Phi-3-Vision, a more sophisticated iteration of their Phi-3 tiny language paradigm. The multimodal model with 4.2 billion parameters is intended for applications like image and chart interpretation. Phi-3-Vision, a member of the Phi-3 family, provides effective visual understanding and is related to Phi-3-mini, Phi-3-small, and Phi-3-medium, which have 3.8 billion, 7 billion, and 14 billion parameters, respectively.
This multimodal model is capable of processing both text and images, making it ideal for mobile device applications. Unlike other image-focused AI models such as OpenAI's DALL-E or Stability AI's Stable Diffusion, Phi-3-Vision does not generate images but analyzes and describes them.
Phi-3-Vision, a 4.2 billion parameter model, is designed to perform general visual reasoning tasks such as interpreting charts or images. Despite its smaller size compared to other AI models, it is claimed to offer efficient image understanding and analysis.
The development of smaller, lightweight AI models like Phi-3 is driven by the growing demand for cost-effective and less compute-intensive AI services. These compact models can power AI features on devices such as smartphones and laptops without overloading computer memory. Microsoft has previously launched other small models including Phi-2 and Orca-Math, a math problem-solving model that reportedly outperforms larger counterparts like Google's Gemini Pro.
See What’s Next in Tech With the Fast Forward Newsletter
Tweets From @varindiamag
Nothing to see here - yet
When they Tweet, their Tweets will show up here.