
"Google announced its widely anticipated Gemini 3 model Tuesday. By many key metrics, it appears to be more capable than the other big generative AI models on the market. In a show of confidence in the performance (and safety) of the new model, Google is making one variant of Gemini-Gemini 3 Pro-available to everyone via the Gemini app starting now. It's also making the same model a part of its core search service for subscribers."
"The new model topped the scores of the much-cited LMArena benchmark, a crowdsourced preference of various top models based on head-to-head responses to identical prompts. In the super-difficult Humanity's Last Exam benchmark test, which measured reasoning and knowledge, the Gemini 3 Pro scored 37.4% compared to GPT-5 Pro's 31.6%. Gemini 3 also topped a range of other benchmarks measuring everything from reasoning to academic knowledge to math to tool use and agent functions."
"Gemini has been a multimodal model from the start, meaning that it can understand and reason about not just language, but images, audio, video, and code-all at the same time. This capability has been steadily improving since the first Gemini, and Gemini 3 reached state-of-the-art performance on the MMMU-Pro benchmark, which measures how well a model handles college-level and professional-level reasoning across text and images. It also topped the Video-MMMU benchmark, which measures the ability to reason over details of video footage."
"For example, the Gemini model might ingest a number of YouTube videos, then create a set of flashcards based on what it learned. Gemini also scored high on its ability to create computer code. That's why it was a good time for the company to launch a new Cursor-like coding agent called Antigravity. Software development has proven to be among the first business functions in which generative AI has had a measurably positive impact."
Google released Gemini 3, a multimodal generative AI model that surpasses many competitors on major benchmarks. One variant, Gemini 3 Pro, is publicly available in the Gemini app and integrated into search for subscribers. Gemini 3 topped LMArena and scored 37.4% on Humanity's Last Exam versus GPT-5 Pro's 31.6%, and it led benchmarks across reasoning, academic knowledge, math, tool use, and agent functions. The model handles text, images, audio, video, and code simultaneously, achieving state-of-the-art results on MMMU-Pro and Video-MMMU. Gemini shows strong code-generation abilities and accompanies a new Cursor-like coding agent called Antigravity to support software development.
Read at Fast Company
Unable to calculate read time
Collection
[
|
...
]