logo
Google's AI charge: How Sergey Brin is taking on the might of OpenAI

Google's AI charge: How Sergey Brin is taking on the might of OpenAI

Mint2 days ago

New Delhi/Mountain View, California: In Mountain View, California, right next to Google's three million square-feet Googleplex headquarters, is a satellite office. While, from the outside, there's nothing seemingly special about it, the building currently houses an elite team of specialist engineers who have been tasked with only one thing: build the best foundational artificial intelligence (AI) model in the world. At the centre of its biggest room sits a man who many in Silicon Valley refer to as a living legend—Sergey Brin, Google's co-founder.
Brin retired in December 2019 but returned to the company last year to lead a light brigade of over 300 engineers, all of whom are charging at OpenAI's GPT models, Google's primary rival in a high stakes battle. OpenAI's GPT models are disrupting the way people search, posing an existential threat to Alphabet Inc., Google's parent company.
Brin is spearheading the development of Gemini, Google's suite of foundational AI models. Gemini's success, or failure, would impact two major areas within Alphabet—Search, and the nascent space of video generation.
For one, Search currently accounts for 56% of Alphabet's annual revenue of $350 billion. Search is also a matter of personal pride for Brin and Larry Page, Google's second founder. Giving up its market dominance in Search means letting go of the duo's legacy—their entire life's work.
Alongside Search, Brin was also concerned about Sora, OpenAI's video generation model. Last year, Google briefly showcased Veo, its video-generating foundational model. However, the market found Veo to be an effort from Google to catch-up with OpenAI.
'This prompted Brin's efforts to create Google Flow this year and launch the AI subscription plans—all a part of his efforts to show that Google, in fact, is still the behemoth as far as Big Tech is concerned," said a senior executive working on the integration of AI in Google's cloud offerings. He didn't want to be identified.
At I/O 2025, an annual developer conference held in May this year, Google launched Flow, a video generation and editing platform that lets users create films with dialogue and background music, without needing any camera, audio and editing setup at all.
A second executive, who also didn't want to be identified, said that much of Google's AI showcase at the conference was driven by what Brin's team has been up to.
'The core task that Brin is leading right now is to prove that Google is not following OpenAI's lead in AI—it is ready to lead innovation for others to follow. Last year, announcements that Google made were all either work in progress, or an iteration of what OpenAI had already showcased. This year, we've largely undone that," the executive, who works with Google's worldwide developer relations teams, said.
A legacy at risk
Much of Google's success, thus far, lies in the 'PageRank' algorithm that made Search the global behemoth that it is today. While the algorithm's patent is owned by Stanford University—Brin's alma mater—he, along with Page, were the ones who invented it.
After failing to sell its algorithm to then-market leader Yahoo twice between 1998 and 2002, Google went on to lead the market globally. In 2021, Yahoo was sold to investment fund Apollo Global Management at $4.88 billion. Alphabet, in 2024, generated $350 billion in annual revenue.
Page, to be sure, is no longer involved with Google's everyday operations, even though he retains a board seat. Instead, Page is focusing on a new AI venture, Dynatomics, which seeks to use generative AI to automate design-led manufacturing of products.
In June 2017, a Stanford University research paper titled 'Attention is all you need', gave birth to the technology behind the transformer model, the fundamental architecture that underpins 'foundational' models. These models, trained on massive troves of data, today crossing trillions, aim to understand, think, calculate and feel like humans. This paper, and the study behind it, was funded by Google.
But Google essentially squandered a technology that it believes it should rightfully lead.
In November 2022, OpenAI—still not well-known back then—introduced ChatGPT, taking the world by storm and causing futurists to predict the doom of human jobs the way we know it today. Others predicted the nascent technology to have spurred into action an 'AI revolution', a seismic shift in the socio-economic balance akin to the industrial revolution of the 18th century.
Alongside OpenAI's shortcut to global stardom, other big tech firms started cashing in on the AI overload. Microsoft was the first to pounce on the opportunity, investing nearly $14 billion in OpenAI and striking various forms of exclusive partnerships. Meta went the open-source way, appearing as a surprise early mover with its Llama family of foundational AI models. By December 2024, Amazon had announced its own family of 'Nova' foundational AI models, even though among Big Tech firms, its direct exposure to AI's algorithmic excellence was the least (Amazon earns its core revenue from e-commerce and cloud services).
Apart from Google, only Apple has so far come off worse. The latter's implementation of AI is yet to see any response of enthusiasm from its customers—and analysts remain sceptical about its ability to keep up with the Big Tech fellows.
Too big, too slow
Analysts state that much of Google's sluggish start in generative AI is attributable to the company's way of functioning. Jayanth N. Kolla, cofounder and partner at consultancy firm Convergence Catalyst, said that at one point, there were concerns internally within senior Google staff that the company was becoming like IBM. 'Too big for its own good, too complacent, and too slow to move on anything," he said.
In 2023, Google shared an internal note following the hype and surge of ChatGPT and OpenAI, asking all its employees to use its internal generative AI platform as much as possible.
'The idea was to maximize the usage hours and mine as much data as possible to bring it up to a certain scale," said a third executive who is with Google's software engineering teams. 'Bard and PaLM (the precursors to Gemini), however, underperformed, which spurred Brin to start taking increasing interest in Google's AI progress," the executive added.
Brin, who turns 52 this August, isn't being strictly shy about his role. At I/O 2025, he made a surprise appearance at a fireside chat with DeepMind chief and Nobel laureate Demis Hassabis.
DeepMind, an AI research laboratory, is a subsidiary of Alphabet.
Speaking about why he came out of retirement, Brin said, 'As a computer scientist, it's a very unique time in history. Honestly, anyone who's a computer scientist should not be retired right now, and be working on AI." He added that he intends to make Gemini 'the world's first AGI, before 2030."
AGI stands for artificial general intelligence, which is loosely defined as an algorithm that mimics the functioning of the human brain, capable of structuring randomized thought, emotion and empathy—qualities that machines lack.
Google showcased more than 16 new products and launches at I/O 2025. The list includes its foundational model's new reasoning capabilities; a 3D video conferencing platform called Google Beam; an always-on version of Gemini Live; a production variant of Project Astra, a multi-modal, all-purpose AI assistant, and Android XR, a new platform for wearable devices.
The headlines, however, were made by Search introducing a new 'AI mode', showcasing for the first time a chat-based interface that changes the way Google's search engine has worked since being incorporated in 1998.
Beating OpenAI
Insiders Mint spoke to said that over the past 12 months, Brin has a single-minded focus—beating OpenAI. A fourth executive working on product management at Google said that the transformer model 'should be rightfully our area of expertise and leadership."
Since 2024, Brin has also been showing up personally at I/O—entering product demos without a prior warning to check on audience feedback.
Executives and analysts believe that Brin's urgency lies in Google's own history. In turn, the executive's return has had a major role in shifting the company's focus—and channeling its focus.
'Sergey has been back since 2023. He's been at work every day focused on AI and Gemini. Another key player is Peter Danenberg who is the godfather of Gemini. In general, the existential threat from Microsoft and Open AI galvanized the entirety of Google to focus on AI," said Ray 'R' Wang, chief executive of US-based tech consulting firm Constellation Research.
Busy Pichai
Brin is bringing unfazed focus to Gemini, Search and Veo, as Sundar Pichai, the CEO of Google and Alphabet, has multiple areas to focus on—lawsuits, global businesses, government relations, cloud, Android and more, the first executive cited above said.
'In the long run, Google foresees its ability to use video generation as a platform to rope in advertisers worldwide, and eventually, establish market dominance in this field," he added.
Pichai, for the longest term, has been viewed as a conservative leader, steering Google's ship with 'one eye on the rear-view mirror," said an analyst who didn't want to be identified.
'For Brin, that's too safe a stance at a time when Silicon Valley is going to war with each other over AI dominance. Plus, Pichai has too much to deal with. Brin's view is that AI today needs undivided attention and he's clearly right, as Google's spate of product launches and share price movement shows," the analyst added.
In the past year, the company's shares are down over 6%, compared to Microsoft's rise of nearly 10%.
While there is no indication that Pichai, who will complete 10 years as the CEO of Google this August (he took over as Alphabet's chief in December 2019), is on his way out, the leadership directives seem to be clearly divided.
Google did not respond to Mint's request for a comment on Brin's recent involvements.
Narrowing gap?
Brin's work may be showing early results.
At a pre-keynote session with journalists during the developer conference, chief executive Pichai said that the Gemini developer platform currently had over seven million developers using its code to create AI applications. This is significant because as of this year, OpenAI's official statistics pegs its outreach at around three million developers.
Earlier this year, at an antitrust lawsuit in a US court, Google conceded that while its developer count is higher than OpenAI's, the latter is still outpacing Google in its monthly active users count. As per filings, OpenAI's ChatGPT platform had over 600 million monthly active users, to Gemini's 350 million. Gemini's numbers, though, are a huge improvement—a year ago, ChatGPT had 400 million monthly active users, in comparison to Gemini's 9 million.
Some analysts do believe that the tide is turning.
'Google is clearly in the lead for AI right now. However, search and ads and mass personalization is about to become more targeted, more actionable, and more intelligent. AI native companies will disrupt existing companies, because intelligence (in business systems) is doubling every seven months—and these AI native companies deliver on exponential efficiency," Constellation's Wang said.
Phil Fersht, chief executive of New York-based tech analysis firm HFS Research, said that Google is 'sitting in an unbelievable position to win the enterprise AI war—if it can get its business model right."
'Net-net, the firm needs to be prepared to cannibalize half of its legacy search business and insert Gemini onto as many enterprises and individual users as possible. It has the resources, talent, and user base to take on OpenAI, Microsoft and Anthropic," he said.
Speed wins
GenAI startups such as OpenAI, Anthropic and Perplexity are known to move fast. They deploy features super quick, reach out to developers and serve a broad variety of AI use cases. Google, in contrast, is viewed to be slower, like Kolla of Convergence Catalyst hinted.
Pichai, speaking with journalists a day ahead of I/O 2025, underlined a new way of working—with speed.
'Typically, we don't make announcements leading up to our big day at I/O each year, but this time it's different. Right now, we're launching products in very frequent intervals, and making technological progress at a rapid pace like never before," he said.
Then, at a post-event chat, Pichai reiterated that Google is now making AI announcements to the world 'within an hour or two" of the DeepMind team showcasing the latest advancements in Gemini.
'In the end, agility and appeal to developers will play the biggest role," said Kashyap Kompella, founder of tech consultancy and research firm RPA2AI Research. 'There's no denying that its rivals are moving fast, and there are clear indications within the industry that Google's AI products are not the first choice for developers and end-users," he added.
The hope is that Brin's startup-style approach, coupled with Google's inherent strength garnered over almost three decades, could be the company's trump card, says Thomas Reuner, principal analyst at UK-based tech consultancy firm PAC.
'Brin might help shore up Google's advertising business in the short term, but its biggest strategic assets are threefold: the vast data assets from the search business, data integration at scale and the unique IP of DeepMind," he said. 'Given the market noise around generative and agentic AI, these assets don't always make the headlines but provide the moat that so many startups are lacking," he added.
Sitting in that satellite office in Mountain View, Brin may be hoping that this moat could firmly establish Gemini, akin to his PageRank moment 29 years ago.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

From Empathy to Innovation: Class 11 Student Creates Tech for the Visually Impaired
From Empathy to Innovation: Class 11 Student Creates Tech for the Visually Impaired

Hans India

time14 minutes ago

  • Hans India

From Empathy to Innovation: Class 11 Student Creates Tech for the Visually Impaired

Seventeen-year-old innovator and accessibility advocate, Ashwat Prasanna, spoke to The Hans India about the journey behind EyeSight—an AI-powered, low-cost smart glasses solution designed to empower India's visually impaired. In this conversation, he shared how empathy, user collaboration, and cutting-edge technology came together to create an inclusive device that aims to reach over 20,000 users by 2026. You're just 17 and already impacting lives—how did your journey into social innovation begin so early? I volunteer at the Premanjali Foundation for the visually impaired. During my time there, I became friends with Charan – he shared my love for math and logic. We would spend a lot of time discussing puzzles and Olympiad questions. But what moved me was when I was told by one of the teachers there not to encourage him a lot since he would never have the opportunities I did – It was very unsettling for me for days – this day and age where we are talking about robot housekeepers and driverless cars, there were still pockets where technology hadn't made its mark – For the next few months, I researched all available accessibility tech, what worked and what was missing – I realized that best results can be achieved if a device was specifically designed for the needs of the visually impaired – navigation, currency reading, scene description etc – that's how the very first version of EyeSight was born – three years ago. Can you share one user testing experience that deeply moved you or reshaped your thinking? More than user testing, I would say that this design was co-created with the users. From the outset, the design and features were influenced by the users' needs and wants. Throughout the version iterations, I got a lot of feedback about what worked and what did not hit the mark. One thing that hit home hard was the affordability. It was easy to get carried away with the latest in technology, but that would be pointless to most of the visually impaired users, as it falls outside their affordability limit. The challenge was to create the best possible version at the lowest cost. How do you plan to make the device sustainable and scalable across India's diverse regions and languages? EyeSight uses the OpenAI API, which has incredible support for India's local languages and even dialects, which currently gives us great reach and localisation in these regions. Additionally, in the future, we plan to fine-tune or train LLMs and AI models to better suit these regions. Another major part of sustainability and scalability is making the device affordable, which has been one of the most significant features so far. Specifically, by making the device 3D printable and using standard parts, it is something that can be assembled by nearly anyone, for everyone. How does EyeSight's offline functionality and ₹1500 pricing truly redefine affordability and accessibility in this space? Compared to other devices with comparable features, which cost upwards of 10-15 thousand rupees, EyeSight is only Rs. 1500, made possible by our choice of design and functionality. Why is this significant? The reality is that for many of the institutions for the visually impaired, more than the features that define access, it's the cost that is the most significant. With EyeSight, thousands more could have access to transformative assistive technology. With a target of 20,000 users by 2026, how do you plan to tackle scale while keeping personalisation and support intact? In the past, the majority of the prototypes have been used in small-scale testing, where they were used individually and not sold to the customer. The pilot phase (set to begin in May) will include loaned units to an institution, by which time the pricing will be finalised. We have received an IB grant of 3000$, which has been very useful in building these late-stage prototypes. Going forward, our first step will be to conduct large-scale user testing and refine the product over the next few months. Based on the testing results, we plan to approach manufacturing units with the refined specs. As far as reaching users is concerned, we are planning to collaborate with schools for the visually impaired in Karnataka. Samarthanam Trust, NAB Karnataka, Mitra Jyoti, and Premanjali Foundation have been of incredible help to us in our creation process. The students in these institutions will be our initial beneficiaries. How did support from programs like IB Global Youth Action Fund and RunWay Incubation shape EyeSight's development? Building the technical product is one thing; taking the product the last mile to reach the market is a whole other thing. As a student, I needed all the help I could get on building EyeSight as a product. RunWay Incubation is a division of PEARL Design Academy that incubates early-stage student ventures such as mine. There I learnt the basic fundamentals of creating a business plan, marketing and fundraising tools. With this foundation, I was able to apply for and acquire the IB Global Youth Action Fund grant of 3000$. This fund, in turn, has helped me build low-fi prototypes and a testable prototype with which I'm doing user testing. How does EyeSight perform offline AI processing on a wearable device without needing constant cloud connectivity? What challenges did you face in optimising performance? Currently, the code implements a combination of models for more detailed online access and quick, essential offline scene inference. This means the basic features of identifying objects, hazards, and safety risks are something that can be possible regardless of an internet connection, and we are working to implement more features to improve offline performance. This is especially significant since many of our users have mentioned that internet connectivity is often patchy in areas where they typically use the product. How do the glasses trigger emergency alerts? Are they gesture-activated or context-based through environmental detection? They need a simple tap gesture on the device before it informs the user and calls emergency services. In future versions, emergencies can be automatically identified using computer vision. What's next for EyeSight after this prototype phase? Are there any new features or partnerships in the works? - First priority is to increase the user-level and field testing for multiple use cases; cooperate with NGOs, and work with the students - From a packaging standpoint, we need to increase the product robustness and reduce the cost of various components; we have identified a hardware partner, and we will accelerate product redesign - We will apply for national and international grants and financial partners for scaling and a large-scale launch

Google rolls out AI tools including NotebookLM for students
Google rolls out AI tools including NotebookLM for students

The Hindu

time20 minutes ago

  • The Hindu

Google rolls out AI tools including NotebookLM for students

Google is enabling access to over 30 AI tools, including NotebookLM, to users under 18 years of age as a part of Google Classroom. Teachers with a Google Workspace account will have a Gemini tab with tools so they can discuss lesson plans, create math problems and write emails. A separate Analytics tab will help teachers track student performance and assignments while showing improvement between them. Subscribers with Education Plus Edition accounts will also be able to access add-ons like Gemini audio lessons. The NotebookLM tool will be made available within the 'coming months,' Google has said so students can generate podcasts on their study material and use interactive study guides. Teachers can build their own custom Gems, mini custom Gemini agents to help students with certain topics. Other tools include Google Vids for generative AI video creation, Google's Class Tools and a new Gemini 'help me create a form' feature will be coming soon to the Google Classroom suite. 'With just a few clicks, educators can create instructional videos that make difficult concepts more digestible. Students can also get creative with Vids and produce their own video book reports and assignments. Vids is integrated with the tools you use every day — like Drive and Classroom — so it's readily accessible,' a blog posted by Google said. The Class Tools feature allows teachers to manage student Chromebooks by sharing videos or other content with them directly.

Google introduces Gemini tool for students and educators: How this AI tool will transform classroom teaching
Google introduces Gemini tool for students and educators: How this AI tool will transform classroom teaching

Time of India

time22 minutes ago

  • Time of India

Google introduces Gemini tool for students and educators: How this AI tool will transform classroom teaching

Gemini ISTE 2025: Gemini for Education launch at ISTE 2025 marks a pivotal moment for classroom AI, with Google unveiling a suite of powerful tools designed for teachers and students worldwide. Built on the cutting‑edge Gemini 2.5 Pro model, this education‑focused version gives schools premium AI features—like interactive diagrams, personalized quiz creation, and automated audio/video overviews—while ensuring data protection and admin control. With more than 30 no‑cost AI tools available through Google Workspace for Education and seamless integration into Classroom, Forms, NotebookLM, and Google Vids, Gemini is redefining how educators plan lessons, engage learners, and evaluate progress. As edtech enters a new AI era, Gemini for Education promises to blend creativity, efficiency, and security—giving schools a smart edge for today and tomorrow. What is Gemini for educators? Gemini for Education is a specialized version of the Gemini app, built on Gemini 2.5 Pro and tailored for academic settings. It offers premium AI models at no extra cost, enhanced data protection, and administrative controls, all of which are included in Google Workspace for Education editions. Premium Access: Default access to cutting-edge AI with higher usage limits than consumer versions. Security & Management: Enterprise-grade protection, admin-managed experience, and Vault integration are standard. All about Gemini app built for education A customized version of the Gemini software, Gemini for Education, was created to meet the particular requirements of the educational sector. Built with the world's top learning model, Gemini 2.5 Pro, Gemini for Education offers enterprise-grade data protection, an admin-managed experience as a core Workspace service, and default access to our premium AI models, soon with much higher limits than what consumers get for free. All these features are included in your Workspace for Education plan at no cost. Because their data is safeguarded by industry-leading security, educational institutions of all sizes can now deploy cutting-edge AI. AI tools for teachers Gemini in Classroom is now available free to all Education edition users. It includes 30+ AI-powered tools to assist educators with tasks like: Generating vocabulary lists with definitions and examples. Drafting rubrics, brainstorming project ideas, translating text, writing stories, tackling misconceptions, and more Teachers can: Refine AI-generated outputs via a chat-like interface. Assign content directly to students within the Classroom. Additionally, custom "Gems" (AI experts) can be created and shared soon, and NotebookLM now supports Audio Overviews—with Video Overviews coming shortly. Paid add-ons offer even more: Google Vids (Veo 3): Create 8-second videos with sound. Gemini in Forms: Generate assessments and surveys from docs/slides/PDFs and receive summarized responses Supporting student learning For students aged 18+, Gemini Canvas enables the generation of personalized quizzes. Soon, this will extend to Younger Students (under 18). Gemini is also introducing interactive visuals like diagrams to aid learning. Teachers can now assign AI-powered Gems and NotebookLM notebooks based on class materials via Google Classroom, PowerSchool, Schoology, and Canvas integrations. Responsible AI & safety measures Google places a high priority on safe, responsible AI: Youth Protections: Under-18 users have stricter content filters and special onboarding featuring AI literacy resources developed with ConnectSafely and FOSI. Fact-Checking: Gemini performs a "double-check" using Google Search for fact-based queries. Data Privacy: No student data is used to train AI models; all educational usage comes with enhanced data protection. Administrators maintain control through admin console settings, access assignment, auditing via Vault, and usage reporting. Gemini also earned the Common Sense Media Privacy Seal.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store