PolyU-led research reveals that sensory and motor inputs help large language models represent complex concepts
A research team led by Prof. Li Ping, Sin Wai Kin Foundation Professor in Humanities and Technology, Dean of the PolyU Faculty of Humanities and Associate Director of the PolyU-Hangzhou Technology and Innovation Research Institute, explored the similarities between large language models and human representations, shedding new light on the extent to which language alone can shape the formation and learning of complex conceptual knowledge.
By exploring the similarities between LLMs and human representations, researchers at The Hong Kong Polytechnic University (PolyU) and their collaborators have shed new light on the extent to which language alone can shape the formation and learning of complex conceptual knowledge. Their findings also revealed how the use of sensory input for grounding or embodiment – connecting abstract with concrete concepts during learning – affects the ability of LLMs to understand complex concepts and form human-like representations. The study, in collaboration with scholars from Ohio State University, Princeton University and City University of New York, was recently published in Nature Human Behaviour.
Led by Prof. LI Ping, Sin Wai Kin Foundation Professor in Humanities and Technology, Dean of the PolyU Faculty of Humanities and Associate Director of the PolyU-Hangzhou Technology and Innovation Research Institute, the research team selected conceptual word ratings produced by state-of-the-art LLMs, namely ChatGPT (GPT-3.5, GPT-4) and Google LLMs (PaLM and Gemini). They compared them with human-generated word ratings of around 4,500 words across non-sensorimotor (e.g., valence, concreteness, imageability), sensory (e.g., visual, olfactory, auditory) and motor domains (e.g., foot/leg, mouth/throat) from the highly reliable and validated Glasgow Norms and Lancaster Norms datasets.
The research team first compared pairs of data from individual humans and individual LLM runs to discover the similarity between word ratings across each dimension in the three domains, using results from human-human pairs as the benchmark. This approach could, for instance, highlight to what extent humans and LLMs agree that certain concepts are more concrete than others. However, such analyses might overlook how multiple dimensions jointly contribute to the overall representation of a word. For example, the word pair 'pasta' and 'roses' might receive equally high olfactory ratings, but 'pasta' is in fact more similar to 'noodles' than to 'roses' when considering appearance and taste. The team therefore conducted representational similarity analysis of each word as a vector along multiple attributes of non-sensorimotor, sensory and motor dimensions for a more complete comparison between humans and LLMs.
The representational similarity analyses revealed that word representations produced by the LLMs were most similar to human representations in the non-sensorimotor domain, less similar for words in sensory domain and most dissimilar for words in motor domain. This highlights LLM limitations in fully capturing humans' conceptual understanding. Non-sensorimotor concepts are understood well but LLMs fall short when representing concepts involving sensory information like visual appearance and taste, and body movement. Motor concepts, which are less described in language and rely heavily on embodied experiences, are even more challenging to LLMs than sensory concepts like colour, which can be learned from textual data.
In light of the findings, the researchers examined whether grounding would improve the LLMs' performance. They compared the performance of more grounded LLMs trained on both language and visual input (GPT-4, Gemini) with that of LLMs trained on language alone (GPT-3.5, PaLM). They discovered that the more grounded models incorporating visual input exhibited a much higher similarity with human representations.
Prof. Li Ping said, 'The availability of both LLMs trained on language alone and those trained on language and visual input, such as images and videos, provides a unique setting for research on how sensory input affects human conceptualisation. Our study exemplifies the potential benefits of multimodal learning, a human ability to simultaneously integrate information from multiple dimensions in the learning and formation of concepts and knowledge in general. Incorporating multimodal information processing in LLMs can potentially lead to a more human-like representation and more efficient human-like performance in LLMs in the future.'
Interestingly, this finding is also consistent with those of previous human studies indicating the representational transfer. Humans acquire object-shape knowledge through both visual and tactile experiences, with seeing and touching objects activating the same regions in human brains. The researchers pointed out that – as in humans – multimodal LLMs may use multiple types of input to merge or transfer representations embedded in a continuous, high-dimensional space. Prof. Li added, 'The smooth, continuous structure of embedding space in LLMs may underlie our observation that knowledge derived from one modality could transfer to other related modalities. This could explain why congenitally blind and normally sighted people can have similar representations in some areas. Current limits in LLMs are clear in this respect'.
Ultimately, the researchers envision a future in which LLMs are equipped with grounded sensory input, for example, through humanoid robotics, allowing them to actively interpret the physical world and act accordingly. Prof. Li said, 'These advances may enable LLMs to fully capture embodied representations that mirror the complexity and richness of human cognition, and a rose in LLM's representation will then be indistinguishable from that of humans.'
Hashtag: #PolyU #HumanCognition #LargeLanguageModels #LLMs #GenerativeAI
The issuer is solely responsible for the content of this announcement.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Bloomberg
2 hours ago
- Bloomberg
Stock Movers: Datadog, FedEx, Synopsys
On this episode of Stock Movers: - Software company Datadog (DDOG) will be added to the S&P 500 index next week, replacing Juniper Networks Inc. after its recent acquisition by Hewlett Packard Enterprise Co. - FedEx (FDX) is double-upgraded to outperform at BNP Paribas Exane, with the broker saying the stock is 'arguably oversold,' expecting the firm's relative operational outperformance vs. rival UPS to continue. UPS is meanwhile upgraded to neutral from underperform. - Synopsys (SNPS) and Cadence Design both gain 6.1% in premarket trading after the US lifted export license requirements for chip design software sales in China, clearing the way for the companies to resume services in the world's second-biggest economy.


Associated Press
2 hours ago
- Associated Press
AEON Partners with Mesh to Unlock Crypto Payments from Major Exchanges and Wallets
HONG KONG, July 3, 2025 /PRNewswire/ -- AEON, the next-generation crypto payment framework, has integrated Mesh, the first global crypto payments network, into its full suite of crypto payment solutions. This partnership allows users to seamlessly spend crypto directly from their exchange accounts and personal wallets across both online and offline merchants within the AEON payment network. With Mesh's API integration, users now can leverage the balances in their accounts on platforms like Coinbase and Binance, or from wallets such as MetaMask, Phantom, and Trust Wallet, using the crypto they already hold as a direct payment method, without transferring funds or converting manually. The connectivity is powered by Mesh's secure technology, which links hundreds of exchanges, wallets, and payment service providers into a single, unified network. This functionality is available across AEON's product lineup, including AI Payment, Online Web3 Payment, and AEON Pay, its Web3 Mobile Payment solution live as Telegram Miniapp and dApp in popular wallets, significantly expanding the payment options available to its users. The partnership unlocks a new level of convenience for crypto holders, enabling them to use their digital assets for a wide array of real-world scenarios. This development is also set to dramatically accelerate crypto adoption, particularly through AEON Pay, the company's Web3 mobile payment solution. AEON Pay is already supported by 20 million merchants across Southeast Asia and is preparing for expansion into emerging global markets, including Africa and Latin America. By providing more payment choices and bridging the gap between digital asset platforms and everyday commerce, the collaboration makes using crypto for daily transactions a practical reality for a massive user base. Founded in 2020, Mesh has built the infrastructure to ensure secure, real-time access to digital asset balances and makes them instantly usable across any connected platform. The partnership between AEON and Mesh represents a pivotal moment in the evolution of digital payments. By turning exchange and wallet balances into spendable capital, this partnership brings unprecedented accessibility, convenience, and interoperability to crypto payments. About Mesh Founded in 2020, Mesh is building the first global crypto payments network, connecting hundreds of exchanges, wallets, and payment service providers (PSPs) to enable seamless crypto payments and stablecoin conversions. By unifying these platforms into a single network, Mesh is pioneering an open, connected, and secure payments ecosystem. For more information, visit LinkedInX About AEON AEON is the next-generation crypto payment framework, built for AI to drive intelligent, automated payments across Web3 and simplify how crypto works in real life. With omni-chain support, AEON's AI Payments system allows intelligent agents to manage and automate transactions, payments, subscriptions, and remittances for users. With offerings like Web3 Mobile Payment that allows users to pay with crypto across 10,000 brands at 20+ million retail merchants in SEA, Africa and Latin America, Online Web3 Payment, Swap Pay, and a growing suite of AI-integrated payment services, AEON aims to create a future of crypto finance that's intelligent, scalable, and borderless, where AI meets real life through seamless, scalable payments for next billion of users. Website | X | Telegram | Medium | AEON Pay View original content: SOURCE AEON

Associated Press
2 hours ago
- Associated Press
Beyond Boundaries: Envision's Record 49-Hour Fire Test Breaks New Ground in Energy Storage Safety
SHANGHAI, July 3, 2025 /PRNewswire/ -- Envision Energy, a global leader in green technology, has successfully completed a groundbreaking large-scale fire test for its smart energy storage system, raising the bar for safety, environmental responsibility, and system resilience. Conducted under full on-site supervision by the Canadian Standards Association (CSA Group), strategic partners, and North American Fire Protection Engineers (FPEs), the test fully complied with the stringent CSA C800 standard – a consensus standard developed with input from insurers, regulators, and industry experts, providing one of the most authoritative international safety benchmarks for energy storage. World's Densest BESS Setup: 5cm Tightest Spacing, Zero Fire Propagation in 49 Hours The test recreated real-world worst-case site conditions by placing four fully charged (100% SOC) 5MWh Envision Smart Energy Storage containers (A, B, C, D) in an ultra-dense back-to-back, side-by-side configuration - just 5 centimeters apart, compared to the typical industry standard spacing of 3 to 5 meters. This setup represents the most energy-dense fire test layout ever attempted, significantly increasing the challenge due to heightened thermal coupling and limited airflow. Importantly, the burn was allowed to proceed with zero intervention. The test pushed the boundaries of the most stringent safety limits - yet even under unprecedented thermal stress and energy concentration, the system showed exceptional resilience and containment. Container A burned for 49 hours and 32 minutes, reaching a peak temperature of 1,297°C. Meanwhile, internal battery temperatures in the adjacent containers B, C, and D remained stable at just 35°C, 44°C, and 43°C respectively - well below our normal operational window. Notably, container A maintained structural integrity throughout the burn, showing no deformation, no collapse, and intact sealing gaskets. Even more impressively, there was zero fire spread or damage to neighboring containers. Units B, C, and D remained unscathed inside and out, with post-test inspections confirming all internal modules were fully operational and undamaged. 3-Hour Ignition Delay - Extended Window for Fire Response To simulate a worst-case ignition scenario, the fire was initiated by applying high-power, synchronized heating to four battery cells within a modified pack, from which the nano-level thermal insulation had been removed. The ignition source was strategically placed at the center of the container cluster - where fire spread risk is greatest in all directions. Under these extreme conditions, ignition took nearly three hours - over 50% longer than typical - offering a critical window for early detection and emergency response. Zero Emission Fire Testing - New Benchmark for Environmental Safety Environmental safety was also a key focus throughout the test. All combustion gases from the burning container A were captured by overhead smoke hoods, analyzed for composition, and fully neutralized through advanced treatment. The entire 49-hour test produced no odors, no emissions, and zero pollution, setting a new industry benchmark for clean and environmentally responsible fire testing in energy storage. AI-Driven Safety & Full Lifecycle Asset Protection Envision's AI-powered Smart Energy Storage System offers superior safety even under extreme conditions. It integrates Envision's proprietary high-safety battery cells within a robust three-layer safety architecture spanning cell, pack, and system levels. Coupled with AI-driven fire detection and active pack-level firefighting, the system isolates thermal events within a single container. To date, Envision's storage systems have been deployed in over 300 projects worldwide with zero safety incidents. This breakthrough fire test proves that even in highly unlikely fire scenarios, Envision's Smart Energy Storage can fully contain risk within a single container. The quantifiable safety results provide insurers with greater confidence to underwrite storage assets, paving the way for broader commercial adoption and accelerating the transition to safe, large-scale energy storage. View original content to download multimedia: SOURCE Envision Energy