top of page

Google introduces the most costly and advanced artificial intelligence model to date, Gemini

Sanjana Mandavia

Dec 6, 2023

The launch of Google's most recent and best AI model, Gemini 1.0, has marked a significant step forward in its journey to artificial intelligence.

Google's new Multimodal Machine Learning Model, named Gemini 1.0, allows you to combine and aggregate information such as text, code, audio, video on a global basis.

Google CEO Sundar Pichai called it the “biggest science and engineering efforts we’ve undertaken as a company.”

Unlike traditional unimodal artificial intelligence systems, which are designed to process multiple types of sensory input in parallel, such as how people's minds perceive the environment, a multimodal system like gemini is designed for processing all kinds of sensory inputs simultaneously. Multimodal artificial intelligence models are trained in the integration and analysis of multimodal data from multiple sources such as text, images, audio or video.

From natural image, audio and video understanding to mathematical reasoning, Gemini Ultra’s performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks used in large language model (LLM) research and development.

With a score of 90.0%, Gemini Ultra is the first model to outperform human experts on MMLU (massive multitask language understanding), which uses a combination of 57 subjects such as math, physics, history, law, medicine and ethics for testing both world knowledge and problem-solving abilities states Demis Hassabis, CEO and Co-Founder of Google DeepMind

This approach enables AI to get a better understanding of data by using several sensors so it can pick up on the same information.

For more advanced explanations, plans, understanding and comprehension, Bard will start using a finetuned version of Gemini Pro from today. Since its introduction, this is the largest upgrade in Bard's history. It will be available in English in more than 170 countries and territories, and we plan to expand to different modalities and support new languages and locations in the near future.

bottom of page