smallbiztechnology_logo (1)

Google releases new models in Gemini AI

3 Min Read
Gemini AI

Google announced updates to its Gemini AI models this week. The company released two new production-ready models, Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002. Google reported improvements in overall quality, especially in math, long context handling, and vision tasks.

https://twitter.com/GoogleDeepMind/status/1840710919044665823

Performance benchmarks show a 7 percent increase overall and a 20 percent improvement in math-related tasks. Google also cut prices for Gemini 1.5 Pro.

https://twitter.com/googlecloud/status/1839741802590441932

Input token costs were reduced by 64 percent and output token costs by 52 percent for prompts under 128,000 tokens.

Rate limits were increased as well. Gemini 1.5 Flash now supports 2,000 requests per minute. Gemini 1.5 Pro can handle 1,000 requests per minute.

These changes aim to make it easier and cheaper for developers to build applications with Gemini. Meta also made AI news this week. The company announced the release of Llama 3.2 on Wednesday.

It’s a major update to Meta’s open-weights AI model lineup. The new release includes vision-capable large language models (LLMs) in 11 billion and 90 billion parameter sizes. It also has lightweight text-only models of 1 billion and 3 billion parameters designed for edge and mobile devices.

Google’s Gemini AI models improve

Meta says the vision models are competitive with leading closed-source models in image recognition and visual understanding tasks. The smaller models outperform similar-sized competitors in various text-based tasks.

See also  Construction industry rapidly adopting digital transformation

The new models support long context windows of up to 128,000 tokens. They are available for free download, with some license restrictions. In other AI news, Google DeepMind announced a big advancement in AI-driven electronic chip design with its AlphaChip technology on Thursday.

The project began in 2020 and has evolved into a reinforcement learning method for designing chip layouts. Google has reportedly used AlphaChip to create “superhuman chip layouts” in the last three generations of its Tensor Processing Units (TPUs). TPUs are chips designed to speed up AI operations.

Google claims AlphaChip can generate high-quality chip layouts in hours. This would take weeks or months of human effort. The company has released a version of AlphaChip on GitHub, sharing the model weights with the public.

AlphaChip’s impact has gone beyond Google. Chip-design companies are adopting and building on the technology for their own chips. This development has sparked new research in AI for chip design.

It could potentially optimize every stage of the process from computer architecture to manufacturing.

Share This Article
Noland Anderson is the driving force behind a cutting-edge technology company at the forefront of digital transformation. As the founder and CEO, Noland combines his deep expertise in tech with a passion for innovation to deliver groundbreaking solutions to clients worldwide.