Latest news with #generalpurposeAI
Yahoo
14 hours ago
- Business
- Yahoo
Coding has emerged as genAI's killer use case. But what if its benefits are a mirage?
Hello and welcome to Eye on AI…In this edition: Meta is going big on data centers…the EU publishes its code of practice for general purpose AI and OpenAI says it will abide by it…the U.K. AI Security Institute calls into question AI 'scheming' research. The big news at the end of last week was that OpenAI's plans to acquire Windsurf, a startup that was making AI software for coding, for $3 billion fell apart. (My Fortune colleague Allie Garfinkle broke that bit of news.) Instead, Google announced that it was hiring Windsurf's CEO Varun Mohan and cofounder Douglas Chen and a clutch of other Windsurf staffers, while also licensing Windsurf's tech—a deal structured similarly to several other big tech-AI startup not-quite-acquihire acquihires, including Meta's recent deal with Scale AI, Google's deal with last year, as well as Microsoft's deal with Inflection and Amazon's with Adept. Bloomberg reported that Google is paying about $2.4 billion for Windsurf's talent and tech, while another AI startup, Cognition, swooped in to buy what was left of Windsurf for an undisclosed sum. Windsurf may have gotten less than OpenAI was offering, but OpenAI's purchase reportedly fell apart after OpenAI and Microsoft couldn't agree on whether Microsoft would have access to Windsurf's increasingly fraught relationship between OpenAI and Microsoft is worth a whole separate story. So too is the structure of these non-acquisition acquihires—which really do seem to blunt any legal challenges, either from regulators or the venture backers of the startups. But today, I want to talk about coding assistants. While a lot of people debate the return on investment from generative AI, the one thing seemingly everyone can agree on is that coding is the one clear killer use case for genAI. Right? I mean, that's why Windsurf was such a hot property and why Anyshphere, the startup behind the popular AI coding assistant Cursor, was recently valued at close to $10 billion. And GitHub Copilot is of course the star of Microsoft's suite of AI tools, with a majority of customers saying they get value out of the product. Well, a trio of papers published this past week complicate this picture. METR, a nonprofit that benchmarks AI models, conducted a randomized control trial involving 16 developers earlier this year to see if using code editor Cursor Pro integrated with Anthropic's Claude Sonnet 3.5 and 3.7 models, actually improved their productivity. METR surveyed the developers before the trial to see if they thought it would make them more efficient and by how much. On average, they estimated that using AI would allow them to complete the assigned coding tasks 24% faster. Then the researchers randomized 246 software coding tasks, either allowing them to be completed with AI or not. Afterwards, the developers were surveyed again on what impact they thought the use of Cursor had actually had on the average time to complete the tasks. They estimated that it made them on average 20% faster. (So maybe not quite as efficient as they had forecast, but still pretty good.) But, and now here's the rub, METR found that when assisted by AI it actually took the coders 19% longer to finish going on here? Well, one issue was that the developers, who were all highly experienced, found that Cursor could not reliably generate code as good as theirs. In fact, they accepted less than 44% of the code-generated responses. And when they did accept them, three-quarters of the developers felt the need to still read over every line of AI-generated code to check it for accuracy, and more than half of the coders made major changes to the Cursor-written code to clean it up. This all took time—on average 9% of the developers time was spent reviewing and cleaning up AI-generated outputs. Many of the tasks in the METR experiment involved large code bases, sometimes consisting of over 100,000 lines of code, and the developers found that sometimes Cursor made strange changes in other parts of this code base that they had to catch and fix. But why did the developers think the AI was making them faster when in fact it was slowing them down? And why, when the researchers followed up with the developers after the experiment ended, did they discover that 69% of the coders were continuing to use Cursor?Some of it seems to be that despite the time it took to edit the Cursor-generated code, the AI assistance did actually ease the cognitive burden for many of the coders. It was mentally easier to fix the AI-generated code than to have to puzzle out the right solution from scratch. So is the perceived ROI from 'vibe coding' itself just vibes? Perhaps. That would actually square with what the Wall Street Journal noted about a different area of genAI use—lawyers using genAI copilots. The newspaper reported that a number of law firms found that given how long it took to fact-check AI-generated legal research, they were not sure lawyers were actually saving any time using the tools. But when they surveyed lawyers, especially junior lawyers, they all reported high satisfaction using the AI copilots and that they felt it made their jobs more a couple of other studies from last week suggest that maybe it all depends on exactly how you use AI coding assistance. A team from Harvard Business School and Microsoft looked at two years of observations of software developers using GitHub Copilot (which is Microsoft product) and found that those using the tool spent more time on coding and less time on project management tasks, in part because GitHub Copilot allowed them to work independently instead of having to use large teams. It also allowed the coders to spend more time exploring possible solutions to coding problems and less time actually implementing the solutions. This too might explain why coders enjoy using these AI tools—because it allows them to spend more time on parts of the job they find intellectually interesting— even if it isn't necessarily about overall time-savings. Finally, let's look at the third study, which is from researchers at Chinese AI startup Modelbest, Chinese universities BUPT and Tsinghua University, and the University of Sydney. They found that while individual AI software development tools often struggled to reliably complete complicated tasks, the results improved markedly when multiple large language models were prompted to each take on a specific role in the software development process and to pose clarifying questions to one another aimed at minimizing hallucinations. They called this architecture 'ChatDev.' So maybe there's a case to be made that the problem with AI coding assistants is how we are using them, not anything wrong with the tech itself? Of course, building teams of AI agents to work in the way ChatDev suggests also uses up a lot more computing power, which gets expensive. So maybe we're still facing that question: is the ROI here a mirage? With that, here's more AI news. Jeremy Before we get to the news, the U.S. paperback edition of my book, Mastering AI: A Survival Guide to Our Superpowered Future, is out from Simon & Schuster. Consider picking up a copy for your bookshelf. Also, if you want to know more about how to use AI to transform your business? Interested in what AI will mean for the fate of companies, and countries? Then join me at the Ritz-Carlton, Millenia in Singapore on July 22 and 23 for Fortune Brainstorm AI Singapore. This year's theme is The Age of Intelligence. We will be joined by leading executives from DBS Bank, Walmart, OpenAI, Arm, Qualcomm, Standard Chartered, Temasek, and our founding partner Accenture, plus many others, along with key government ministers from Singapore and the region, top academics, investors and analysts. We will dive deep into the latest on AI agents, examine the data center build out in Asia, examine how to create AI systems that produce business value, and talk about how to ensure AI is deployed responsibly and safely. You can apply to attend here and, as loyal Eye on AI readers, I'm able to offer complimentary tickets to the event. Just use the discount code when you The essay above was written and edited by Fortune staff. The news items below were selected by the newsletter author, created using AI, and then edited and fact-checked. This story was originally featured on Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data


France 24
6 days ago
- Business
- France 24
EU unveils recommendations to rein in powerful AI models
Brussels has come under fierce pressure to delay enforcing its landmark AI law as obligations for complex models known as general-purpose AI -- systems that have a vast range of functions -- kick in from August 2. The law entered into force last year but its different obligations will apply gradually. But as the EU pivots to bolstering its competitivity and catching up with the United States and China, European tech firms and some US Big Tech want Brussels to slow down. The European Commission, the bloc's digital watchdog, has pushed back against a delay. The EU's executive arm has now published a code of practice for such systems prepared by independent experts with input from others including model providers themselves. In the code, the experts recommend practical measures such as excluding known piracy websites from the data models use. The code applies to general-purpose AI models, such as Google's Gemini, Meta's Llama and X's Grok -- the tech billionaire Elon Musk's chatbot that has come under fire this week for antisemitic comments. Under the law, developers of such models must give details about what content they used -- like text or images -- to train their systems and comply with EU copyright law. The code was due to be published in May. EU officials reject claims that it had been watered down in the past few months due to industry pressure. Corporate Europe Observatory and Lobby Control in April had accused Big Tech of "heavily" influencing the process "to successfully weaken the code". The code will need endorsement by EU states before companies can voluntarily sign up to it. Businesses that sign the code "will benefit from a reduced administrative burden and increased legal certainty compared to providers that prove compliance in other ways", the commission said in a statement. Nearly 50 of Europe's biggest companies including France's Airbus, Dutch tech giant ASML and Germany's Lufthansa and Mercedes-Benz last week urged a two-year pause. The companies' CEOs in a letter accused the EU's complex rules of putting at risk the 27-country bloc's AI ambitions and the development of European champions. The EU will be able to enforce the rules for general-purpose AI models a year from August 2 for new models, while existing models will have until August 2027 to comply.