Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

最新消息最新消息
0

Google has recently announced the release of two upgraded production-ready Gemini 1.5 models, Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002, designed to provide enhanced performance, speed, and cost-effectiveness. These models, which are based on significant improvements to the original Gemini 1.5 model, are expected to revolutionize the field of AI in various applications.

Enhanced Performance and Cost Efficiency

The new models, Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002, are part of Google’s ongoing efforts to refine and optimize its AI models. According to the official announcement, these models show significant improvements in areas such as mathematics, code generation, long-text context, and visual tasks. Specifically, the updated models have seen performance boosts in challenging benchmarks like MMLU-Pro and mathematics and hidden math datasets.

Pricing and Speed Improvements

One of the most notable changes is the reduction in pricing for the Gemini 1.5 models. For inputs and outputs with prompt tokens less than 128K, the price of the Gemini 1.5 Pro model has been slashed by more than 50%. Additionally, the Gemini 1.5 Flash model now has doubled processing rates, with the Gemini 1.5 Pro model seeing a 3-fold increase in output speed and a 3-fold reduction in latency.

Developer Access and Availability

Developers can now access the latest models through Google AI Studio and the Gemini API. For large enterprises and Google Cloud customers, the models can also be obtained via Vertex AI. The improvements are particularly significant in areas such as mathematics, long text context, and visual tasks, making the models more versatile and robust for a wide range of applications.

Improved Response and Cost Reduction

The updated models have also seen improvements in response quality and efficiency. The new models are designed to reduce the number of instances where they provide no help or incorrect answers, with more responses being helpful and concise. For use cases such as summarization, Q&A, and information extraction, the default output length has been reduced by 5-20% compared to previous models. This reduction in output length not only saves costs but also enhances the usability of the models.

Migration and Further Improvements

Google has also made it easier for developers to migrate to the latest Gemini 1.5 Pro and 1.5 Flash versions. For example, the input token price for the Gemini 1.5 Pro model has been reduced by 64%, and the output token price has been cut by 52%, with incremental cache token prices also reduced by 64%. These changes are expected to further reduce the overall cost of using Gemini for applications.

Enhanced Rate Limits

To facilitate easier application development, Google has increased the rate limits for the Gemini 1.5 Flash model to 2000 RPM and for the Gemini 1.5 Pro model to 1000 RPM, up from 1000 and 360 respectively. The company expects to continue raising the rate limits for the Gemini API in the coming weeks, allowing developers to build more applications with the models.

Performance Enhancements

The latest models have also seen significant performance enhancements. The Gemini 1.5 Flash-8B experimental update, released on September 24, shows marked improvements in text and multimodal use cases. This model is now accessible via Google AI Studio and the Gemini API.

Conclusion

The release of the upgraded Gemini 1.5 models marks a significant milestone in Google’s AI development efforts. With improved performance, cost efficiency, and enhanced usability, these models are poised to become indispensable tools for developers and businesses looking to leverage advanced AI technology. As Google continues to refine and expand its AI offerings, the future looks bright for both the tech industry and its users.


>>> Read more <<<

Views: 0

0

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注