OpenAI announced that its new flagship model GPT-4o mini is officially available on Azure AI! This powerful model has excellent text processing capabilities, and will also support image, audio and video processing in the future, bringing users a more comprehensive AI experience. GPT-4o mini scored as high as 82% in the multi-task language understanding test, which is significantly better than GPT-3.5 Turbo, and is cheaper and extremely cost-effective. It also has an expanded 128K context window and improved multi-language capabilities, allowing users to handle more complex and larger tasks with ease.
OpenAI announced today that its new flagship model GPT-4o mini has been launched on Azure AI at the same time, supporting text processing capabilities, and will launch image, audio and video functions in the future.
GPT-4o mini is significantly smarter than GPT-3.5Turbo—scoring 82% on the massive multi-task language understanding (MMLU) measure, compared to 70% for GPT-3.5Turbo—and more than 60% cheaper. 1 This model provides an expanded 128K context window and integrates GPT-4o’s improved multi-language capabilities. Azure OpenAI Studio Playground offers a free trial of GPT-4o mini.
Microsoft Azure AI brings default security, expanded data residency and service availability upgrades to GPT-4o mini. Customers can expect to experience improved performance and functionality on Azure AI, especially for streaming scenarios such as assistants, code interpreters, and retrieval.
Azure AI announced global pay-as-you-go and maximum throughput limits for GPT-4o mini. Customers now have the flexibility to pay based on the resources they consume, while traffic is routed globally to provide higher throughput and static control over where data is stored. The global pay-as-you-go deployment option will provide a throughput of 15 million tokens per minute (TPM), providing 99.99% availability for GPT-4o mini and the same industry rates as OpenAI.
GPT-4o mini will be available on Azure AI this month and will be available in the Batch service. Batch delivers high-throughput jobs within 24 hours at a 50% discount by using off-peak capacity. This is only possible if Microsoft runs on Azure AI, which enables Microsoft Azure AI to provide off-peak capacity to customers.
Microsoft Azure AI will also release the fine-tuning function of GPT-4o mini this month, allowing customers to further customize the model according to specific use cases and scenarios. Following last month's update to token-based training billing, Microsoft Azure AI has reduced hosting fees by 43%. Combined with its low inference price, this makes Azure OpenAI service fine-tuned deployment the most cost-effective offering for customers with production workloads.
Plan highlights:
⭐ GPT-4o mini is launched on Azure AI, supports text processing functions, is extremely fast and has comprehensive functions
⭐ The new model is smarter than its predecessor, more than 60% cheaper, offers a wider context window and multi-language capabilities
⭐ Azure AI provides global pay-as-you-go deployment options for GPT-4o mini, providing high throughput and 99.99% availability
In short, the release of GPT-4o mini on Azure AI marks a new stage of more powerful, more economical, and easier-to-use AI technology, bringing more opportunities to developers and enterprises. Its high performance, high availability and flexible pricing model will undoubtedly become a powerful engine for the development of AI applications in the future.