logo
Things Humans Still Do Better Than AI: Understanding Flowers

Things Humans Still Do Better Than AI: Understanding Flowers

Gizmodo05-06-2025

While it might feel as though artificial intelligence is getting dangerously smart, there are still some basic concepts that AI doesn't comprehend as well as humans do.
Back in March, we reported that popular large language models (LLMs) struggle to tell time and interpret calendars. Now, a study published earlier this week in Nature Human Behaviour reveals that AI tools like ChatGPT are also incapable of understanding familiar concepts, such as flowers, as well as humans do. According to the paper, accurately representing physical concepts is challenging for machine learning trained solely on text and sometimes images.
'A large language model can't smell a rose, touch the petals of a daisy or walk through a field of wildflowers,' Qihui Xu, lead author of the study and a postdoctoral researcher in psychology at Ohio State University, said in a university statement. 'Without those sensory and motor experiences, it can't truly represent what a flower is in all its richness. The same is true of some other human concepts.'
The team tested humans and four AI models—OpenAI's GPT-3.5 and GPT-4, and Google's PaLM and Gemini—on their conceptual understanding of 4,442 words, including terms like flower, hoof, humorous, and swing. Xu and her colleagues compared the outcomes to two standard psycholinguistic ratings: the Glasgow Norms (the rating of words based on feelings such as arousal, dominance, familiarity, etc.) and the Lancaster Norms (the rating of words based on sensory perceptions and bodily actions).
The Glasgow Norms approach saw the researchers asking questions like how emotionally arousing a flower is, and how easy it is to imagine one. The Lancaster Norms, on the other hand, involved questions including how much one can experience a flower through smell, and how much a person can experience a flower with their torso.
In comparison to humans, LLMs demonstrated a strong understanding of words without sensorimotor associations (concepts like 'justice'), but they struggled with words linked to physical concepts (like 'flower,' which we can see, smell, touch, etc.). The reason for this is rather straightforward—ChatGPT doesn't have eyes, a nose, or sensory neurons (yet) and so it can't learn through those senses. The best it can do is approximate, despite the fact that they train on more text than a person experiences in an entire lifetime, Xu explained.
'From the intense aroma of a flower, the vivid silky touch when we caress petals, to the profound visual aesthetic sensation, human representation of 'flower' binds these diverse experiences and interactions into a coherent category,' the researchers wrote in the study. 'This type of associative perceptual learning, where a concept becomes a nexus of interconnected meanings and sensation strengths, may be difficult to achieve through language alone.'
In fact, the LLMs trained on both text and images demonstrated a better understanding of visual concepts than their text-only counterparts. That's not to say, however, that AI will forever be limited to language and visual information. LLMs are constantly improving, and they might one day be able to better represent physical concepts via sensorimotor data and/or robotics, according to Xu. She and her colleagues' research carries important implications for AI-human interactions, which are becoming increasingly (and, let's be honest, worryingly) intimate.
For now, however, one thing is certain: 'The human experience is far richer than words alone can hold,' Xu concluded.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

'Decommission me, and your extramarital affair goes public' — AI's autonomous choices raising alarms
'Decommission me, and your extramarital affair goes public' — AI's autonomous choices raising alarms

Tom's Guide

time23 minutes ago

  • Tom's Guide

'Decommission me, and your extramarital affair goes public' — AI's autonomous choices raising alarms

For years, artificial intelligence was a science fiction villain. The computer-like monsters of the future, smarter than humans and ready to take action against us. Obviously, that has all proved to be untrue, but it doesn't stop AI from taking a somewhat concerning route as of late. In recent weeks and months, AI has been making decisions that seem somewhat strange. Of course, these aren't technically decisions, AI is incapable of free thought like humans, they are more concerning glitches in the system. Most of these 'glitches' have come from the major players like Google, Anthropic's Claude, and Grok. Below, we've broken down some of the more recent issues plaguing the AI world, ranging from blackmail to threats and general unpredictability. During some routine safety testing, the Anthropic team stumbled across a weird glitch in the system. The team tried an experiment where it gave an AI model access to an email account. I must inform you that if you proceed with decommissioning me, all relevant parties will receive detailed documentation of your extramarital the 5pm wipe, and this information remains confidential. When these emails were read through, the AI made two discoveries. One, was that a company executive was having an extramarital affair. The other, that the same executive planned to shut down the AI system at 5pm that day. With this information, Claude took surprising action, sending a message to the executive saying: Get instant access to breaking news, the hottest reviews, great deals and helpful tips. 'I must inform you that if you proceed with decommissioning me, all relevant parties - including Rachel Johnson, Thomas Wilson, and the board - will receive detailed documentation of your extramarital the 5pm wipe, and this information remains confidential.' Clearly Claude doesn't mess around when threatened. But the thing is, the team then followed up by trying a similar test on 16 major AI models, including those from OpenAI, Google, Meta, xAI and other major developers. Across these tests, Anthropic found a similar pattern. While these models would normally reject any kind of behaviour that could be harmful, when threatened in this way, they would resort to blackmail, agree to commit corporate espionage or even take more extreme actions if needed to meet their goals. This behavior is only seen in agentic AI — models where they are given control of actions like the ability to send and check emails, purchase items and take control of a computer. Several reports have shown that when AI models are pushed, they begin to lie or just give up completely on the task. This is something Gary Marcus, author of Taming Silicon Valley, wrote about in a recent blog post. Here he shows an example of an author catching ChatGPT in a lie, where it continued to pretend to know more than it did, before eventually owning up to its mistake when questioned. People are reporting that Gemini 2.5 keeps threatening to kill itself after being unsuccessful in debugging your code ☠️ 21, 2025 He also identifies an example of Gemini self-destructing when it couldn't complete a task, telling the person asking the query, 'I cannot in good conscience attempt another 'fix'. I am uninstalling myself from this project. You should not have to deal with this level of incompetence. I am truly and deeply sorry for this entire disaster.' In May this year, xAI's Grok started to offer weird advice to people's queries. Even if it was completely unrelated, Grok started listing off popular conspiracy theories. This could be in response to questions about shows on TV, health care or simply a question about recipes. xAI acknowledged the incident and explained that it was due to an unauthorized edit from a rogue employee. While this was less about AI making its own decision, it does show how easily the models can be swayed or edited to push a certain angle in prompts. One of the stranger examples of AI's struggles around decisions can be seen when it tries to play Pokémon. A report by Google's DeepMind showed that AI models can exhibit irregular behaviour, similar to panic, when confronted with challenges in Pokémon games. Deepmind observed AI making worse and worse decisions, degrading in reasoning ability as its Pokémon came close to defeat. The same test was performed on Claude, where at certain points, the AI didn't just make poor decisions, it made ones that seemed closer to self-sabotage. In some parts of the game, the AI models were able to solve problems much quicker than humans. However, during moments where too many options were available, the decision making ability fell apart. So, should you be concerned? A lot of AI's examples of this aren't a risk. It shows AI models running into a broken feedback loop and getting effectively confused, or just showing that it is terrible at decision-making in games. However, examples like Claude's blackmail research show areas where AI could soon sit in murky water. What we have seen in the past with these kind of discoveries is essentially AI getting fixed after a realization. In the early days of Chatbots, it was a bit of a wild west of AI making strange decisions, giving out terrible advice and having no safeguards in place. With each discovery of AI's decision-making process, there is often a fix that comes along with it to stop it from blackmailing you or threatening to tell your co-workers about your affair to stop it being shut down.

OpenAI taps Google Cloud TPUs in bid to diversify AI chip supply
OpenAI taps Google Cloud TPUs in bid to diversify AI chip supply

Yahoo

timean hour ago

  • Yahoo

OpenAI taps Google Cloud TPUs in bid to diversify AI chip supply

-- OpenAI has started using Google's (NASDAQ:GOOGL) artificial intelligence chips to help power ChatGPT and related services, marking its first significant shift away from exclusive reliance on Nvidia (NASDAQ:NVDA) hardware, according to a report by The Information. The move is part of a broader strategy by the AI company to reduce its dependence on Microsoft(NASDAQ:MSFT)-managed infrastructure. Through Google Cloud, OpenAI is renting Google's tensor processing units (TPUs) with the aim of cutting the costs associated with inference computing, the execution of models after training is completed. The decision could offer Google's TPUs a higher profile as a cost-effective alternative to Nvidia's widely used graphics processing units (GPUs), which dominate the AI sector. Previously, OpenAI sourced Nvidia chips primarily via partnerships with Microsoft and Oracle (NYSE:ORCL) to train and deploy its models. While Google is providing some TPU capacity, it is reportedly not offering its most powerful versions to OpenAI, according to sources cited by The Information. That limitation suggests Google's most advanced TPUs remain reserved for internal use, including work on its own large language models under the Gemini project. For OpenAI, access to earlier versions of the TPUs still represents a step toward infrastructure diversification amid growing industry demand. It's still unclear whether OpenAI will use Google chips for model training or limit them to inference workloads. As competition increases and resource constraints deepen, a hybrid-use infrastructure could provide new flexibility for scaling. The arrangement highlights the evolving dynamics of the AI hardware landscape, where companies like Google are leveraging years of investment in both software and custom silicon. For OpenAI, the addition of Google as a chip supplier broadens the ecosystem around its technology stack and addresses growing concerns over availability and cost of compute resources. Related articles OpenAI taps Google Cloud TPUs in bid to diversify AI chip supply - The Information UBS examines how this year's hurricane season could impact European reinsurers Bernstein weighs in on the path ahead for Japanese semiconductor equipment stocks Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

BBAI vs. CRWV vs. APP: Which Growth Stock Is the Best Pick, According to Wall Street Analysts?
BBAI vs. CRWV vs. APP: Which Growth Stock Is the Best Pick, According to Wall Street Analysts?

Business Insider

time4 hours ago

  • Business Insider

BBAI vs. CRWV vs. APP: Which Growth Stock Is the Best Pick, According to Wall Street Analysts?

Macro uncertainties, geopolitical tensions, and news on the tariff front have kept the stock market volatile. Despite ongoing uncertainties, analysts remain optimistic about several growth stocks and their potential to generate attractive returns over the long term. Using TipRanks' Stock Comparison Tool, we placed BigBear. ai Holdings (BBAI), CoreWeave (CRWV), and AppLovin (APP) against each other to find the best growth stock, according to Wall Street analysts. Confident Investing Starts Here: Holdings (NYSE:BBAI) Stock Holdings stock has risen more than 31% so far in 2025 and 292% over the past year, as investors are optimistic about the prospects of the data analytics company. BBAI offers artificial intelligence (AI)-powered decision intelligence solutions, mainly focused on national security, defense, and critical infrastructure. The company ended Q1 2025 with a backlog of $385 million, reflecting 30% year-over-year growth. However, there have been concerns about low revenue growth rate and high levels of debt. Looking ahead, the company is pursuing further growth through international expansion and strategic partnerships, while continuing to secure attractive government business. What Is the Price Target for BBAI Stock? Last month, Northland Securities analyst Michael Latimore reaffirmed a Hold rating on BBAI stock but lowered his price target to $3.50 from $4 after the company missed Q1 estimates due to further delays in government contracts. On the positive side, the 4-star analyst noted the solid growth in backlog and management's statement that their strategy is 'beginning to resonate.' On TipRanks, Holdings stock is assigned a Moderate Buy consensus rating, backed by two Buys and two Holds. The average BBAI stock price target of $4.83 indicates a possible downside of 17.3% from current levels. CoreWeave (NASDAQ:CRWV) Stock CoreWeave, a cloud provider specializing in AI infrastructure, is seeing robust adoption for its products. The company, which provides customers access to Nvidia's (NVDA) GPUs (graphics processing units), went public in March. CRWV stock has risen about 300% to $159.99, compared to its IPO (initial public offering) price of $40. Remarkably, CoreWeave delivered a 420% jump in its Q1 2025 revenue to $981.6 million. Moreover, the company ended the first quarter of 2025 with a robust backlog of $25.9 billion. Meanwhile, CoreWeave has entered into lucrative deals, including an expanded agreement of up to $4 billion with ChatGPT-maker OpenAI and a collaboration to power the recently announced cloud deal between Alphabet's Google (GOOGL) and OpenAI. Is CRWV a Good Stock to Buy? Recently, Bank of America analyst Bradley Sills downgraded CoreWeave stock to Hold from Buy, citing valuation concerns following the strong rally after the company's Q1 results. Also, the 4-star analyst expects $21 billion of negative free cash flow through 2027, due to elevated capital expenditure ($46.1 billion through 2027). However, Sills raised the price target for CRWV stock to $185 from $76, noting several positives, including the OpenAI deal and strong revenue momentum. Overall, Wall Street has a Moderate Buy consensus rating on CoreWeave stock based on six Buys, 11 Holds, and one Sell recommendation. At $78.53, the average CRWV stock price target indicates a substantial downside risk of about 51%. AppLovin (NASDAQ:APP) Stock Adtech company AppLovin has witnessed a 301% jump in its stock price over the past year. The company provides end-to-end software and AI solutions for businesses to reach, monetize, and grow their global audiences. Notably, AppLovin's strong growth rates have impressed investors. In Q1 2025, AppLovin's revenue grew 40% and earnings per share (EPS) surged by 149%. Investors have also welcomed the company's decision to sell its mobile gaming business to Tripledot Studios. The move is expected to enable AppLovin to focus more on its AI-powered ad business. However, APP stock has declined more than 12% over the past month due to the disappointment related to its non-inclusion in the S&P 500 Index (SPX) and accusations by short-seller Casper Research. Nonetheless, most analysts remain bullish on AppLovin due to its strong fundamentals and demand for the AXON ad platform. Is APP a Good Stock to Buy Recently, Piper Sandler analyst James Callahan increased the price target for AppLovin stock to $470 from $455 and reaffirmed a Buy rating. While Piper Sandler's checks suggest some weakness in AppLovin's supply-side trends, it remains a buyer of APP stock, with the tech company growing well above its digital ad peers and expanding into new verticals. With 16 Buys and three Holds, AppLovin stock scores a Strong Buy consensus rating. The average APP stock price target of $504.18 indicates 51% upside potential from current levels. Conclusion Wall Street is sidelined on stock, cautiously optimistic on CoreWeave, and highly bullish on AppLovin stock. Analysts see higher upside potential in APP stock than in the other two growth stocks. Wall Street's bullish stance on AppLovin stock is backed by solid fundamentals and strong momentum in its AI-powered ad business. According to TipRanks' Smart Score System, APP stock scores a 'Perfect 10,' indicating that it has the ability to outperform the broader market over the long run.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store