logo
Nvidia-backed Perplexity launches AI-powered browser to take on Google Chrome

Nvidia-backed Perplexity launches AI-powered browser to take on Google Chrome

The Hindu10-07-2025
Nvidia-backed Perplexity AI said on Wednesday it has launched Comet, a new web browser with AI-powered search capabilities, as the startup looks to challenge the dominance of market leader Alphabet's Google Chrome.
The launch marks Perplexity AI's entry into the competitive browser market, aiming to replace traditional navigation with agentic AI that can think, act, and decide on behalf of users.
Google Chrome held a commanding 68% share of the global browser market in June, according to StatCounter, cementing its position as the world's most widely used browser—far ahead of Safari, Microsoft Edge, and Firefox.
Comet enables users to ask questions, perform tasks, and conduct research in a single, unified interface. The browser integrates a built-in assistant that can compare products, summarise content, book meetings, and transform complex workflows into simple, conversational experiences.
Comet is currently available to subscribers who pay $200 per month for Perplexity Max, with broader access rolling out via invite over the summer.
Backed by high-profile investors like Jeff Bezos, SoftBank, and Nvidia , Perplexity is leveraging the launch of Comet to directly challenge incumbents such as Google's Chrome and Microsoft's Edge, while also exploring new revenue streams in advertising and e-commerce.
Last year, OpenAI added a search engine to ChatGPT and recently made the service available to all of its users. Google also introduced an AI-powered search feature, called AI Overviews, last May.
Comet stores data locally and avoids model training on personal information—a move likely to appeal to privacy-conscious users.
However, the company has come under criticism from media organizations such as News Corp-owned outlets, Forbes and Wired, and Wall Street Journal parent Dow Jones for using their content without consent or compensation. Perplexity has responded by launching a publisher partnership program to collaborate with news outlets.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Meet Lumo, the new AI chatbot that protects user privacy
Meet Lumo, the new AI chatbot that protects user privacy

Indian Express

time2 hours ago

  • Indian Express

Meet Lumo, the new AI chatbot that protects user privacy

Proton, the company that introduced the encrypted email service Proton Mail, has now unveiled an AI chatbot with focus on user privacy. Named Lumo, the chatbot can generate code, write email, summarise documents, and much more. Proton has dubbed its AI chatbot as an alternative to ChatGPT, Gemini, Copilot, etc. The AI chatbot preserves user privacy while storing data locally on users' devices. Lumo is powered by several open-source large language models that run on Proton's servers in Europe, including Mistral's Nemo, Mistral Small 3, Nvidia's OpenHands 32B, and the Allen Institute for AI's OLMO 2 32B model. Lumo can field requests through different models depending on which is better suited for a query. The company claims that the new chatbot will protect information with 'zero-access' encryption, which grants the user an encryption key that allows them exclusive access to their data. This encryption key will block third parties and even Proton from accessing the user content, meaning the company will not be sharing any personal information. Proton has reportedly used Transport Layer Security (TLS) encryption for data transmission and 'asymmetrically' encrypts prompts, allowing only the Lumo GPU servers to decrypt them. When it comes to features, Ghost mode ensures that your active chat sessions are not saved, not even on local devices. With the Web search feature, Lumo can look up recent or new information on the internet to add to its current knowledge. It can also understand and analyse your files, but does not keep a record of them. Lastly, integration with Proton Drive makes it simple to add end-to-end encrypted files from your Proton Drive to your Lumo chats. The chatbot comes with internet search, however, it is disabled by default to ensure privacy. Once enabled, Lumo will deploy privacy-friendly search engines to provide responses to user queries. It can analyse uploaded files, but it does not store any of the data. Proton Drive files, which are meant to remain end-to-end encrypted while communicating with the chatbot, can also be linked by users to Lumo. The chatbot comes in both a free and premium version. Those without an account with Lumo or Proton, will be able to ask 25 queries per week. They will not be able to access chat histories. On the other hand, users with a free account can ask up to 100 questions per week. Lumo Plus plan is priced at $12.99 a month and comes with unlimited chats, an extended encrypted chat history, and more.

No hiring in India: Could Trump's new push disrupt tech jobs?
No hiring in India: Could Trump's new push disrupt tech jobs?

India Today

time2 hours ago

  • India Today

No hiring in India: Could Trump's new push disrupt tech jobs?

Donald Trump has called on Google, Apple, and other US tech giants to stop hiring in India and prioritise Americans instead. The statement, though not yet a policy, has triggered concern across India's white-collar job sector, especially among engineering and management graduates aiming for roles in global tech enforced, the impact could be sharp: from IIT placement trends to mid-level tech hiring, from GCC operations to India's growing dependence on global companies for high-skilled and Wisdom Hatch Founder Akshat Shrivastava said on X that the chances for Indian jobseekers are LET'S UNDERSTAND HOW THE HIRING WORKS American tech companies don't just hire from India. Some also build in India, including the Google, Apple, and Microsoft, all have large operations here, often in the form of Global Capability Centres (GCCs).These giants maintain a presence in India where professionals are at the task of building products, writing code, testing machines, overseeing global employs around 10,000 people in India. Microsoft has over 18,000 employees here. These are for operations across the has around 5,000 direct employees in India, and thousands more in its supplier network and development of these jobs are not for call centres or support. The core development roles, white-collar positions, are in play that attract graduates from IITs, NITs, IIITs, and even Tier-2 colleges.A pronounced number of these roles are built for India but are part of global projects. And yes, some top-tier students are hired directly to go to the US exactly the pipe Trump wants to close."There's no denying that companies like Google and Apple have long symbolised the 'dream job' for many Indian students, and rightfully so. Indian talent has been a major contributor to the global tech revolution, not just as employees but as leaders," says Neelakantha Bhanu, Founder and CEO of Bhanzu, and title holder of 'World's Fastest Human Calculator'."However, if such hiring freezes become a reality, it will be a wake-up call, not in fear, but in perspective," he says."The world is changing, and so are opportunities. India today is not just a source of talent, but a builder of global products," Bhanu adds. WHAT HAPPENS TO IIT AND IIM PLACEMENTS?Every December, the buzz begins: placement season. But behind the success stories, there's a truth not often told -- many students don't land dream IITians, despite the brand tag, end up in jobs that pay Rs 8-10 lakh per annum or less. Not because they aren't brilliant. But because not everyone gets picked by Google, Microsoft, or a US -- based startup with a fancy from IITs from 2023-2024 show that even in top IITs, 20-25% of students were still unplaced at the end of the obtained through RTI requests filed by IIT Kanpur alumnus Dheeraj Singh shows that nearly 8,000 students, around 38% across 23 IITs, remain unplaced in the year IIMs, especially the older ones, place most students in India-based roles, consulting, banking, and management BIG PICTURE: INDIA'S WHITE-COLLAR WALL COULD CRACKTrump's statement comes at a time when India is producing more engineers than it can absorb. Private colleges, deemed universities, and even Tier-1 institutes are churning out thousands of tech graduates each year. But demand has 51.25% of graduates amongst the graduates in India are considered employable, highlighting persistent gaps in vocational training and skill development, as per the Economic Survey turns out nearly 15 lakh engineering graduates each year, yet only 10-15% among them find employment, as noted in a report by has kept things afloat over the past decade is the globalisation of Indian tech talent. US -- Mbased hiring, remote work, global team integration, and GCC expansion have created a top 10% of tech graduates, those who would go abroad or work on US -- facing roles from India, may have to compete in a shrinking domestic market. And this creates a domino effect down the ladder."If US tech companies stop hiring from India, it'll cost them more than us. India has long been their strongest talent pool, from engineers to CEOs. Some students may miss out on overseas roles, but fewer than 2% of IIT graduates go abroad now," says Nishant Chandra, Co-founder, Newton School."Most choose to stay and lead from India. This shift could actually benefit us by putting focus on skills over pedigree," he SMALLER STARTUPS FOLLOW SUIT?Possibly. If the bigger players hit pause, mid-sized companies may rethink their hiring plans too, especially those who build for US clients or rely on US venture capital. And in India's startup ecosystem, perception drives could delay hiring cycles, reduce internship opportunities, and force more candidates to settle for lower Group Founder Ankur Agarwal, a top executive search firm, sums it up: 'These Trump rules, if enforced strictly, will definitely impact placements in IITs as the top US companies recruit quite a bit from these campuses for their US -- based tech development. IIMs are usually used to hire for India roles only, so they will not be impacted.""The real impact will be felt by the GCCs, though, which have become an important recruiter for top quartile tech talent. However, the actual impact will depend on how strictly companies comply with this directive and whether it becomes formal policy, as the US still faces significant tech talent shortages that make complete elimination of overseas hiring challenging," he NEXT?Nobody knows if Trump's statement will become law. But it's already a signal."Our institutions, our ecosystems, and our ambitions are ready. And as someone who chose to stay and build here, I can say that there's never been a more exciting time to be in India. We're not just producing global talent anymore. We're producing global solutions," says time to prepare is now, not just with coding skills, but with adaptability, global exposure, and maybe even a Plan B that doesn't rely on a Silicon Valley zip code.- Ends advertisement

The new chips designed to solve AI's energy problem
The new chips designed to solve AI's energy problem

Mint

time3 hours ago

  • Mint

The new chips designed to solve AI's energy problem

'I can't wrap my head around it," says Andrew Wee, who has been a Silicon Valley data-center and hardware guy for 30 years. The 'it" that has him so befuddled—irate, even—is the projected power demands of future AI supercomputers, the ones that are supposed to power humanity's great leap forward. Wee held senior roles at Apple and Meta, and is now head of hardware for cloud provider Cloudflare. He believes the current growth in energy required for AI—which the World Economic Forum estimates will be 50% a year through 2030—is unsustainable. 'We need to find technical solutions, policy solutions and other solutions that solve this collectively," he says. To that end, Wee's team at Cloudflare is testing a radical new kind of microchip, from a startup founded in 2023, called Positron, which has just announced a fresh round of $51.6 million in investment. These chips have the potential to be much more energy efficient than ones from industry leader Nvidia at the all-important task of inference, which is the process by which AI responses are generated from user prompts. While Nvidia chips will continue to be used to train AI for the foreseeable future, more efficient inference could collectively save companies tens of billions of dollars, and a commensurate amount of energy. There are at least a dozen chip startups all battling to sell cloud-computing providers the custom-built inference chips of the future. Then there are the well-funded, multiyear efforts by Google, Amazon and Microsoft to build inference-focused chips to power their own internal AI tools, and to sell to others through their cloud services. The intensity of these efforts, and the scale of the cumulative investment in them, show just how desperate every tech giant—along with many startups—is to provide AI to consumers and businesses without paying the 'Nvidia tax." That's Nvidia's approximately 60% gross margin, the price of buying the company's hardware. Nvidia is very aware of the growing importance of inference and concerns about AI's appetite for energy, says Dion Harris, a senior director at Nvidia who sells the company's biggest customers on the promise of its latest AI hardware. Nvidia's latest Blackwell systems are between 25 and 30 times as efficient at inference, per watt of energy pumped into them, as the previous generation, he adds. To accomplish their goals, makers of novel AI chips are using a strategy that has worked time and again: They are redesigning their chips, from the ground up, expressly for the new class of tasks that is suddenly so important in computing. In the past, that was graphics, and that's how Nvidia built its fortune. Only later did it become apparent graphics chips could be repurposed for AI, but arguably it's never been a perfect fit. Jonathan Ross is chief executive of chip startup Groq, and previously headed Google's AI chip development program. He says he founded Groq (no relation to Elon Musk's xAI chatbot) because he believed there was a fundamentally different way of designing chips—solely to run today's AI models. Groq claims its chips can deliver AI much faster than Nvidia's best chips, and for between one-third and one-sixth as much power as Nvidia's. This is due to their unique design, which has memory embedded in them, rather than being separate. While the specifics of how Groq's chips perform depends on any number of factors, the company's claim that it can deliver inference at a lower cost than is possible with Nvidia's systems is credible, says Jordan Nanos, an analyst at SemiAnalysis who spent a decade working for Hewlett Packard Enterprise. Positron is taking a different approach to delivering inference more quickly. The company, which has already delivered chips to customers including Cloudflare, has created a simplified chip with a narrower range of abilities, in order to perform those tasks more quickly. The company's latest funding round came from Valor Equity Partners, Atreides Management and DFJ Growth, and brings the total amount of investment in the company to $75 million. Positron's next-generation system will compete with Nvidia's next-generation system, known as Vera Rubin. Based on Nvidia's road map, Positron's chips will have two to three times better performance per dollar, and three to six times better performance per unit of electricity pumped into them, says Positron CEO Mitesh Agrawal. Competitors' claims about beating Nvidia at inference often don't reflect all of the things customers take into account when choosing hardware, says Harris. Flexibility matters, and what companies do with their AI chips can change as new models and use cases become popular. Nvidia's customers 'are not necessarily persuaded by the more niche applications of inference," he adds. Cloudflare's initial tests of Positron's chips were encouraging enough to convince Wee to put them into the company's data centers for more long-term tests, which are continuing. It's something that only one other chip startup's hardware has warranted, he says. 'If they do deliver the advertised metrics, we will open the spigot and allow them to deploy in much larger numbers globally," he adds. By commoditizing AI hardware, and allowing Nvidia's customers to switch to more-efficient systems, the forces of competition might bend the curve of future AI power demand, says Wee. 'There is so much FOMO right now, but eventually, I think reason will catch up with reality," he says. One truism of the history of computing is that whenever hardware engineers figure out how to do something faster or more efficiently, coders—and consumers—figure out how to use all of the new performance gains, and then some. Mark Lohmeyer is vice president of AI and computing infrastructure for Google Cloud, where he provides both Google's own custom AI chips, and Nvidia's, to Google and its cloud customers. He says that consumer and business adoption of new, more demanding AI models means that no matter how much more efficiently his team can deliver AI, there is no end in sight to growth in demand for it. Like nearly all other big AI providers, Google is making efforts to find radical new ways to produce energy to feed that AI—including both nuclear power and fusion. The bottom line: While new chips might help individual companies deliver AI more efficiently, the industry as a whole remains on track to consume ever more energy. As a recent report from Anthropic notes, that means energy production, not data centers and chips, could be the real bottleneck for future development of AI. Write to Christopher Mims at

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store