Latest news with #languagemodels


Forbes
07-07-2025
- Entertainment
- Forbes
Deep Research And More From New AI Models
Young vocalist with a microphone singing on stage Did you hate these annoying songs? Check out what new deep research AI tools can do. It's all well and good to tell people to get on ChatGPT, or Claude, or Gemini, and get their hands dirty using an AI model – but some people want more detail. They want to know how to use the best new tools in a more granular way. Deep research is one of the newest and most impressive functions of today's large language models. But most people haven't heard of it yet, let alone used it. Ethan Mollick is someone I associate with the vanguard of AI research. With former MIT connections, Mollick represents a leading voice in this field. So I was reading his latest blog post about different models, and how to use them. Mollick breaks down the most useful models for Claud, ChatGPT and Google in the following categories: 'chat,' 'work' and 'hard problems.' Claude, he notes, is generally better for privacy, where you need to select for this in ChatGPT, and may compromise some functionality with Google. Then Mollick embarks on a description of deep research capabilities that are baked into some of the newest models, but not yet available for free. This is so new that it's still paywalled and rationed for paying users. The Promise of Deep Research Functionality Deep research is available from OpenAI, and from Google, and from Anthropic, too. When you look at the unveiling of the research function, you see a lot of reports around February. So how do you access this kind of result? I found a recent post about this from the 'OpenAI cookbook' in June, where the poster talks about downloading a Python SDK, and API work. It turns out, though, that you can select 'deep research' in the prompt section of ChatGPT 4.5, if you are a paying Plus or Pro user, and get your deep research report that way. Asking for Deep Research from ChatGPT 4.5 I wanted to try this for myself, so based on what I was experiencing in the moment, I centered on compiling a report of the most annoying songs from 1990 through 2025. My query: 'please create a report on the most annoying music from 1990 until 2025, segmented annually.' Simple, right? First, ChatGPT asked me to clarify a few basic questions: 'To create a detailed report on the 'most annoying music' from 1990 to 2025, could you please clarify a few things: And I answered: '1. all of the above 2. US-based 3. multiple songs per year 4. Yes.' 'Thanks!' GPT responded. 'I'll compile a U.S.-focused report covering the most widely criticized, overplayed, or culturally mocked songs from 1990 to 2025. This will include multiple songs per year, along with context explaining why they were considered annoying—drawing from music critics, listener polls, memes, radio overplay complaints, and public opinion.' Then the report started. Indeed, it took more than 10 minutes to compile, with CoT displays like 'searching 'Rico Suave': worst song' and 'looking for Macarena.' Check Out GPT's Chain of Thought Actually, here are some of the interesting CoT notes the system provided while doing research on dozens of sites (it spent a lot of time reading the Guardian!) I'm mapping out recent headlines and thought pieces, like those from Global News and the Guardian, all noting "Baby Shark" as the most annoying song. Hm, that's interesting. I'm preparing the list of most annoying songs from 1993-1995, considering "What's Up," "Cotton Eye Joe," "Macarena," and "Wonderwall." However, I'm thinking through the inclusion of other notable tracks like "Rico Suave" for earlier years. (There's that 'Rico Suave' song again.) I'm mapping out the hits and misses of 2001-2006, with 2005's viral sensations like "Crazy Frog" and "Photograph" by Nickelback taking center stage. Hm, that's interesting—these tracks seem to evoke strong reactions. (Note how the model is actually 'taking notes' while researching, giving us little bits of personality along with the actual meat of the report.) Here's a funny comparison: I'm comparing the impact of Nickelback's "Rockstar" in 2006 to Soulja Boy's "Crank That" in 2007. The former was widely criticized, while the latter was loved by many, raising questions about musical taste over time. A research-based note: I'm digging through songs by genre and year, considering widespread opinion and specific examples to select the most frequently labeled annoying tracks. I'm evaluating hits like Meghan Trainor's "All About That Bass" (2014) and Tyga's "Stimulated" (2015) for their ubiquity and frequency of being labeled annoying. Progressing through last five years gradually. I'm digging through songs by genre and year, considering widespread opinion and specific examples to select the most frequently labeled annoying tracks. I'm evaluating hits like Meghan Trainor's "All About That Bass" (2014) and Tyga's "Stimulated" (2015) for their ubiquity and frequency of being labeled annoying. Progressing through last five years gradually. And then there was this: 'Noticing the challenge of time limitations, I'm figuring out how to balance inclusion of pivotal examples against the need for conciseness.' So the model is actually trying to triage its work. This is next-level stuff. Finally, GPT appeared to wrap up, dotting its I's and crossing its Ts: 'I'm checking to see if the 'Achy Breaky' song contains any references to anthropomorphic sharks.' Eventually, I got the full report. It ranges dozens of pages, and provides great coverage of over a quarter century of horrible, terrible music. A few standouts, with years: 1990 1994 1997 Fast-forward a bit: 2013 2023 Some Thoughts on Voice Mode Let's move back to the blog post: after delving into the above stuff, Mollick also covers the verbal capabilities of new systems that really change the interface. As he notes, you can talk to these models while you're in the car, or on a walk, or anywhere that you have the space and time. I think verbal communication is going to finally make LLM responses 'real' to us. And that's all I have to say about that – for now. I think all of this is deeply informing to people who want to be practical about LLM use. The deep research function in particular is something people will be learning about as time goes on. How many of us will figure it out before something new comes along that makes it obsolete? In any case, this was an excellent guide to how to make the most of AI in today's environment.
Yahoo
27-06-2025
- Business
- Yahoo
Hugging Face Co-Founder Challenges AI Optimists: 'Models Can't Ask Original Scientific Questions'
Thomas Wolf, co-founder and chief science officer at Hugging Face, has cast doubt on the belief that current artificial intelligence systems will lead to major scientific breakthroughs. Wolf told Fortune that today's large language models, or LLMs, excel at providing answers but fall short when it comes to formulating original questions. 'In science, asking the question is the hard part,' he said. 'Once the question is asked, often the answer is quite obvious, but the tough part is really asking the question, and models are very bad at asking great questions.' Don't Miss: GoSun's breakthrough rooftop EV charger already has 2,000+ units reserved — become an investor in this $41.3M clean energy brand today. Invest early in CancerVax's breakthrough tech aiming to disrupt a $231B market. Back a bold new approach to cancer treatment with high-growth potential. Wolf's comments were in response to a blog post by Anthropic CEO Dario Amodei, who argues that artificial intelligence could compress a century's worth of scientific breakthroughs into just a few years. Wolf said he initially found the post compelling but became skeptical after rereading. 'It was saying AI is going to solve cancer, and it's going to solve mental health problems—it's going to even bring peace into the world. But then I read it again and realized there's something that sounds very wrong about it, and I don't believe that,' he told Fortune. San Francisco-based Anthropic is backed by tech giants, including Inc. (NASDAQ:AMZN) and Alphabet Inc. (NASDAQ:GOOG, GOOGL)), and is also known for its Claude family of AI models. For Wolf, the core issue lies in how LLMs are trained. In another blog post, Wolf argues that today's AI systems are built to predict likely outcomes, act as "yes-men on servers," capable of mimicking human responses but incapable of challenging assumptions or generating original ideas. "To create an Einstein in a data center, we don't just need a system that knows all the answers, but rather one that can ask questions nobody else has thought of or dared to ask," Wolf wrote. He cited that real scientific progress often comes from paradigm shifts—like Copernicus proposing heliocentrism or the invention of CRISPR-based gene editing—rather than from answering existing questions. Trending: This Jeff Bezos-backed startup will allow you to become a landlord in just 10 minutes, with minimum investments as low as $100. Wolf also questioned how AI performance is measured today. In his blog post, he pointed to benchmarks like Humanity's Last Exam or Frontier Math, which test how well AI models can answer complex but well-defined questions. "These are exactly the kinds of exams where I excelled," Wolf wrote, referencing his academic background. "But real scientific breakthroughs come not from answering known questions, but from asking challenging new ones and questioning previous ideas." He argued that AI needs to demonstrate the ability to challenge its training data, take counterfactual approaches, and identify new research directions from incomplete information. Using the board game Go as an analogy, Wolf said the landmark 2016 victory of DeepMind's AlphaGo over world champions made headlines but was not revolutionary. "Move 37, while impressive, is still essentially a straight-A student answer to the question posed by the rules of the game of Go," he wrote in his blog. "An Einstein-level breakthrough in Go would involve inventing the rules of Go itself." Hugging Face is a prominent open-source platform in the AI community, known for its collaborative development of open-source machine learning models and tools. The company is backed by investors including Sequoia Capital and Lux Capital, and it plays a leading role in developing transparent and accessible AI systems. Wolf concluded that while current models are useful as assistants, true scientific progress requires a different kind of intelligence—one that can formulate disruptive questions rather than repeat what is already known. See Next: $100k in assets? Maximize your retirement and cut down on taxes: Schedule your free call with a financial advisor to start your financial journey – no cost, no obligation. Warren Buffett once said, "If you don't find a way to make money while you sleep, you will work until you die." Here's how you can earn passive income with just $100. UNLOCKED: 5 NEW TRADES EVERY WEEK. Click now to get top trade ideas daily, plus unlimited access to cutting-edge tools and strategies to gain an edge in the markets. Get the latest stock analysis from Benzinga? (AMZN): Free Stock Analysis Report ALPHABET (GOOG): Free Stock Analysis Report This article Hugging Face Co-Founder Challenges AI Optimists: 'Models Can't Ask Original Scientific Questions' originally appeared on © 2025 Benzinga does not provide investment advice. All rights reserved. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data
Yahoo
27-06-2025
- Business
- Yahoo
Hugging Face Co-Founder Challenges AI Optimists: 'Models Can't Ask Original Scientific Questions'
Thomas Wolf, co-founder and chief science officer at Hugging Face, has cast doubt on the belief that current artificial intelligence systems will lead to major scientific breakthroughs. Wolf told Fortune that today's large language models, or LLMs, excel at providing answers but fall short when it comes to formulating original questions. 'In science, asking the question is the hard part,' he said. 'Once the question is asked, often the answer is quite obvious, but the tough part is really asking the question, and models are very bad at asking great questions.' Don't Miss: GoSun's breakthrough rooftop EV charger already has 2,000+ units reserved — become an investor in this $41.3M clean energy brand today. Invest early in CancerVax's breakthrough tech aiming to disrupt a $231B market. Back a bold new approach to cancer treatment with high-growth potential. Wolf's comments were in response to a blog post by Anthropic CEO Dario Amodei, who argues that artificial intelligence could compress a century's worth of scientific breakthroughs into just a few years. Wolf said he initially found the post compelling but became skeptical after rereading. 'It was saying AI is going to solve cancer, and it's going to solve mental health problems—it's going to even bring peace into the world. But then I read it again and realized there's something that sounds very wrong about it, and I don't believe that,' he told Fortune. San Francisco-based Anthropic is backed by tech giants, including Inc. (NASDAQ:AMZN) and Alphabet Inc. (NASDAQ:GOOG, GOOGL)), and is also known for its Claude family of AI models. For Wolf, the core issue lies in how LLMs are trained. In another blog post, Wolf argues that today's AI systems are built to predict likely outcomes, act as "yes-men on servers," capable of mimicking human responses but incapable of challenging assumptions or generating original ideas. "To create an Einstein in a data center, we don't just need a system that knows all the answers, but rather one that can ask questions nobody else has thought of or dared to ask," Wolf wrote. He cited that real scientific progress often comes from paradigm shifts—like Copernicus proposing heliocentrism or the invention of CRISPR-based gene editing—rather than from answering existing questions. Trending: This Jeff Bezos-backed startup will allow you to become a landlord in just 10 minutes, with minimum investments as low as $100. Wolf also questioned how AI performance is measured today. In his blog post, he pointed to benchmarks like Humanity's Last Exam or Frontier Math, which test how well AI models can answer complex but well-defined questions. "These are exactly the kinds of exams where I excelled," Wolf wrote, referencing his academic background. "But real scientific breakthroughs come not from answering known questions, but from asking challenging new ones and questioning previous ideas." He argued that AI needs to demonstrate the ability to challenge its training data, take counterfactual approaches, and identify new research directions from incomplete information. Using the board game Go as an analogy, Wolf said the landmark 2016 victory of DeepMind's AlphaGo over world champions made headlines but was not revolutionary. "Move 37, while impressive, is still essentially a straight-A student answer to the question posed by the rules of the game of Go," he wrote in his blog. "An Einstein-level breakthrough in Go would involve inventing the rules of Go itself." Hugging Face is a prominent open-source platform in the AI community, known for its collaborative development of open-source machine learning models and tools. The company is backed by investors including Sequoia Capital and Lux Capital, and it plays a leading role in developing transparent and accessible AI systems. Wolf concluded that while current models are useful as assistants, true scientific progress requires a different kind of intelligence—one that can formulate disruptive questions rather than repeat what is already known. See Next: $100k in assets? Maximize your retirement and cut down on taxes: Schedule your free call with a financial advisor to start your financial journey – no cost, no obligation. Warren Buffett once said, "If you don't find a way to make money while you sleep, you will work until you die." Here's how you can earn passive income with just $100. UNLOCKED: 5 NEW TRADES EVERY WEEK. Click now to get top trade ideas daily, plus unlimited access to cutting-edge tools and strategies to gain an edge in the markets. Get the latest stock analysis from Benzinga? (AMZN): Free Stock Analysis Report ALPHABET (GOOG): Free Stock Analysis Report This article Hugging Face Co-Founder Challenges AI Optimists: 'Models Can't Ask Original Scientific Questions' originally appeared on © 2025 Benzinga does not provide investment advice. All rights reserved. Se produjo un error al recuperar la información Inicia sesión para acceder a tu portafolio Se produjo un error al recuperar la información Se produjo un error al recuperar la información Se produjo un error al recuperar la información Se produjo un error al recuperar la información
Yahoo
22-06-2025
- Science
- Yahoo
Scientists Just Found Something Unbelievably Grim About Pollution Generated by AI
Tech companies are hellbent on pushing out ever more advanced artificial intelligence models — but there appears to be a grim cost to that progress. In a new study in the science journal Frontiers in Communication, German researchers found that large language models (LLM) that provide more accurate answers use exponentially more energy — and hence produce more carbon — than their simpler and lower-performing peers. In other words, the findings are a grim sign of things to come for the environmental impacts of the AI industry: the more accurate a model is, the higher its toll on the climate. "Everyone knows that as you increase model size, typically models become more capable, use more electricity and have more emissions," Allen Institute for AI researcher Jesse Dodge, who didn't work on the German research but has conducted similar analysis of his own, told the New York Times. The team examined 14 open source LLMs — they were unable to access the inner workings of commercial offerings like OpenAI's ChatGPT or Anthropic's Claude — of various sizes and fed them 500 multiple choice questions plus 500 "free-response questions." Crunching the numbers, the researchers found that big, more accurate models such as DeepSeek produce the most carbon compared to chatbots with smaller digital brains. So-called "reasoning" chatbots, which break problems down into steps in their attempts to solve them, also produced markedly more emissions than their simpler brethren. There were occasional LLMs that bucked the trend — Cogito 70B achieved slightly higher accuracy than DeepSeek, but with a modestly smaller carbon footprint, for instance — but the overall pattern was stark: the more reliable an AI's outputs, the greater its environmental harm. "We don't always need the biggest, most heavily trained model, to answer simple questions," Maximilian Dauner, a German doctoral student and lead author of the paper, told the NYT. "Smaller models are also capable of doing specific things well. The goal should be to pick the right model for the right task." That brings up an interesting point: do we really need AI in everything? When you go on Google, those annoying AI summaries pop up, no doubt generating pollution for a result that you never asked for in the first place. Each individual query might not count for much, but when you add them all up, the effects on the climate could be immense. OpenAI CEO Sam Altman, for example, recently enthused that a "significant fraction" of the Earth's total power production should eventually go to AI. More on AI: CEOs Using AI to Terrorize Their Employees