In the early morning of May 22, Microsoft Build 2024 Developers Conference was held in Seattle, the United States. According to Satya Nadella, CEO and Chairman of Microsoft, the latest flagship model GPT-4o developed by OpeanAI is now available in Azure AI Studio as an API. This multimodal model integrates text, image and audio processing capabilities, bringing a new generative and conversational AI experience.
In addition, Phi-3-vision, a new multimodal model in the Phi-3 column AI Small Language Model (SLM) developed by Microsoft, has been launched in Azure. The Phi-3 model is powerful, cost-effective and optimized for personal devices. Phi-3-vision provides the function of inputting images and text and receiving text responses. For example, users can ask questions about charts or open-ended questions about specific images.
Developers can test these cutting-edge models in Azure AI Playground, and can start to build and customize using the models in Azure AI Studio.