PolyU-led research reveals that sensory and motor inputs help large language models represent complex concepts
A research team led by Prof. Li Ping, Sin Wai Kin Foundation Professor in Humanities and Technology, Dean of the PolyU Faculty of Humanities and Associate Director of the PolyU-Hangzhou Technology and Innovation Research Institute, explored the similarities between large language models and human representations, shedding new light on the extent to which language alone can shape the formation and learning of complex conceptual knowledge.
By exploring the similarities between LLMs and human representations, researchers at The Hong Kong Polytechnic University (PolyU) and their collaborators have shed new light on the extent to which language alone can shape the formation and learning of complex conceptual knowledge. Their findings also revealed how the use of sensory input for grounding or embodiment – connecting abstract with concrete concepts during learning – affects the ability of LLMs to understand complex concepts and form human-like representations. The study, in collaboration with scholars from Ohio State University, Princeton University and City University of New York, was recently published in Nature Human Behaviour.
Led by Prof. LI Ping, Sin Wai Kin Foundation Professor in Humanities and Technology, Dean of the PolyU Faculty of Humanities and Associate Director of the PolyU-Hangzhou Technology and Innovation Research Institute, the research team selected conceptual word ratings produced by state-of-the-art LLMs, namely ChatGPT (GPT-3.5, GPT-4) and Google LLMs (PaLM and Gemini). They compared them with human-generated word ratings of around 4,500 words across non-sensorimotor (e.g., valence, concreteness, imageability), sensory (e.g., visual, olfactory, auditory) and motor domains (e.g., foot/leg, mouth/throat) from the highly reliable and validated Glasgow Norms and Lancaster Norms datasets.
The research team first compared pairs of data from individual humans and individual LLM runs to discover the similarity between word ratings across each dimension in the three domains, using results from human-human pairs as the benchmark. This approach could, for instance, highlight to what extent humans and LLMs agree that certain concepts are more concrete than others. However, such analyses might overlook how multiple dimensions jointly contribute to the overall representation of a word. For example, the word pair 'pasta' and 'roses' might receive equally high olfactory ratings, but 'pasta' is in fact more similar to 'noodles' than to 'roses' when considering appearance and taste. The team therefore conducted representational similarity analysis of each word as a vector along multiple attributes of non-sensorimotor, sensory and motor dimensions for a more complete comparison between humans and LLMs.
The representational similarity analyses revealed that word representations produced by the LLMs were most similar to human representations in the non-sensorimotor domain, less similar for words in sensory domain and most dissimilar for words in motor domain. This highlights LLM limitations in fully capturing humans' conceptual understanding. Non-sensorimotor concepts are understood well but LLMs fall short when representing concepts involving sensory information like visual appearance and taste, and body movement. Motor concepts, which are less described in language and rely heavily on embodied experiences, are even more challenging to LLMs than sensory concepts like colour, which can be learned from textual data.
In light of the findings, the researchers examined whether grounding would improve the LLMs' performance. They compared the performance of more grounded LLMs trained on both language and visual input (GPT-4, Gemini) with that of LLMs trained on language alone (GPT-3.5, PaLM). They discovered that the more grounded models incorporating visual input exhibited a much higher similarity with human representations.
Prof. Li Ping said, 'The availability of both LLMs trained on language alone and those trained on language and visual input, such as images and videos, provides a unique setting for research on how sensory input affects human conceptualisation. Our study exemplifies the potential benefits of multimodal learning, a human ability to simultaneously integrate information from multiple dimensions in the learning and formation of concepts and knowledge in general. Incorporating multimodal information processing in LLMs can potentially lead to a more human-like representation and more efficient human-like performance in LLMs in the future.'
Interestingly, this finding is also consistent with those of previous human studies indicating the representational transfer. Humans acquire object-shape knowledge through both visual and tactile experiences, with seeing and touching objects activating the same regions in human brains. The researchers pointed out that – as in humans – multimodal LLMs may use multiple types of input to merge or transfer representations embedded in a continuous, high-dimensional space. Prof. Li added, 'The smooth, continuous structure of embedding space in LLMs may underlie our observation that knowledge derived from one modality could transfer to other related modalities. This could explain why congenitally blind and normally sighted people can have similar representations in some areas. Current limits in LLMs are clear in this respect'.
Ultimately, the researchers envision a future in which LLMs are equipped with grounded sensory input, for example, through humanoid robotics, allowing them to actively interpret the physical world and act accordingly. Prof. Li said, 'These advances may enable LLMs to fully capture embodied representations that mirror the complexity and richness of human cognition, and a rose in LLM's representation will then be indistinguishable from that of humans.'
Hashtag: #PolyU #HumanCognition #LargeLanguageModels #LLMs #GenerativeAI
The issuer is solely responsible for the content of this announcement.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
an hour ago
- Yahoo
Alibaba Aims for Capital Efficiency: Can the Banma Spin-Off Succeed?
Alibaba BABA is betting big on capital efficiency with its proposal to spin off Banma Network Technology and list it independently in Hong Kong. The move shifts funding for Banma's fast-scaling auto-software business to public markets, reducing the company's direct cash burden while preserving upside through a minority stake. Under the plan, Alibaba's holding will drop from about 44.72% to just over 30%, subject to regulatory approvals, including the spin-off addresses a key financial priority, redeploying capital toward higher-return areas like cloud, AI and share buybacks. Alibaba reported a robust net cash position of RMB 366.4 billion as of March 31, and generated RMB 73.9 billion in free cash flow in fiscal 2025, while returning nearly $16.5 billion to its shareholders through repurchases and dividends. Freeing itself from Banma's funding needs ensures that capital can be allocated to this core growth and return spin-off appears to be a win-win deal for both companies. It unlocks greater valuation potential for Banma as an independent mobility software leader while allowing Alibaba to streamline operations and focus on core spin-off's success will largely depend on execution, navigating regulatory hurdles, securing an attractive IPO price and converting OEM alliances into consistent revenue growth. Should these efforts succeed, Alibaba could boost capital returns, concentrate on high-growth core operations and unlock significant shareholder value by turning Banma into a tradable asset. Alibaba Faces Intensifying Cloud Competition Amazon's AMZN-owned Amazon Web Services (AWS) leads the global cloud market, especially in North America and Europe, while Alibaba Cloud dominates Asia-Pacific, particularly China and is expanding globally. Amazon's AWS is known for its mature, feature-rich services, covering compute, storage, databases, networking and analytics. Although Amazon's pricing is generally higher, it offers flexible models suited for diverse workloads. Amazon's extensive capabilities and global reach make it a formidable competitor for Alibaba GOOGL-owned Google Cloud Platform offers extensive global reach and seamless integration with the Google ecosystem. Google Cloud is highly regarded for its AI and ML strengths, including tools like Vertex AI and TensorFlow. While Alibaba Cloud provides AI services tailored for practical business applications such as recommendation engines and customer service bots, Google Cloud's superior data storage, management, and advanced AI capabilities give it a competitive edge. Google Cloud continues to expand its influence worldwide. BABA's Share Price Performance, Valuation & Estimates BABA shares have gained 40.9% in the year-to-date period, outperforming the Zacks Internet – Commerce industry and the Zacks Retail-Wholesale sector's growth of 11.6% and 8.6%, respectively. BABA's YTD Price Performance Image Source: Zacks Investment Research From a valuation standpoint, BABA stock is currently trading at a forward 12-month Price/Earnings ratio of 12.75X compared with the industry's 25.17X. BABA has a Value Score of C. BABA's Valuation Image Source: Zacks Investment Research The Zacks Consensus Estimate for first-quarter fiscal 2026 earnings is pegged at $2.13 per share, which remains unchanged over the past 30 days, indicating a 5.75% year-over-year decline. Image Source: Zacks Investment Research Alibaba currently carries a Zacks Rank #5 (Strong Sell).You can see the complete list of today's Zacks #1 Rank (Strong Buy) stocks here. Want the latest recommendations from Zacks Investment Research? Today, you can download 7 Best Stocks for the Next 30 Days. Click to get this free report Inc. (AMZN) : Free Stock Analysis Report Alphabet Inc. (GOOGL) : Free Stock Analysis Report Alibaba Group Holding Limited (BABA) : Free Stock Analysis Report This article originally published on Zacks Investment Research ( Zacks Investment Research Sign in to access your portfolio
Yahoo
an hour ago
- Yahoo
3 Reasons I'm Buying Nvidia's Stock Like There's No Tomorrow
Key Points Next year's data center buildouts will reach record levels. Nvidia's H20 chips bound for China will boost revenue The stock isn't as expensive as some think. 10 stocks we like better than Nvidia › Nvidia (NASDAQ: NVDA) is the world's largest company, which may give some investors pause before investing in it. A common mindset is: "If it's already the biggest company in the world, can it get any bigger?" From my analysis, yes, it can get much bigger, and it will continue to be a winning investment from here. I have three reasons Nvidia is still an excellent buy at its current price tag, and investors could be glad they picked up shares three years down the road. 1. Data center capital expenditures are still rising Nvidia's primary products, graphics processing units (GPUs), have been in huge demand thanks to the AI race. They have become the computing unit of choice for training and running these models, and that dominance doesn't look to be faltering. GPUs can process multiple calculations in parallel, leading to impressive performance. And they can be connected in clusters within a data center to create unparalleled computing power for AI applications. Luckily for investors, Nvidia's largest clients are starting to offer guidance for 2026 capital expenditures (capex). A high percentage of this revenue will be used for data centers, with a hefty chunk going toward filing them with Nvidia GPUs. Even though 2025 was a record-setting year for capex, 2026 is expected to exceed these levels. The company also expects this to continue rising through at least 2028. During its 2025 GTC event, management cited a third-party projection that estimated global data center capex at $400 billion during 2024. That's expected to rise to $1 trillion by 2028, which will be a huge growth boost for Nvidia. This primary growth engine for the company is far from done, which bodes well for its future. 2. Nvidia's China business will return shortly In April, the Trump administration revoked Nvidia's export license for its H20 chips, which were specifically designed to meet prior export regulations. This created a significant hole in the company's business, but that hole could be plugged shortly. It reapplied for its export license with assurances from the government that it will be approved. However, there's one concession that the company may need to make. The government plans to charge Nvidia a 15% export tax on these GPUs, which will eat into its profit margins. Even so, these chip sales to China will be a huge boost to its business, even if they aren't as profitable as they once were. In the second quarter, its projected revenue growth would have been 77% instead of the guidance figure of 50% if H20 chip sales were allowed. This boost could be coming as soon as the third quarter, providing another crucial growth lever. 3. The stock isn't as expensive as it seems Nvidia's rise hasn't been all business-related. Its stock has easily become the most popular in the market, which has caused its valuation to skyrocket. At 42 times forward earnings, it is far from cheap. However, the company's growth is more than enough to offset that. If Wall Street analysts' projections for $5.91 in earnings per share (EPS) in fiscal 2027 (ending January 2027) turn out to be right, that indicates the stock trades for 31 times 2027 earnings, which is a far more reasonable price tag. However, the average analyst only expects 27% revenue growth in fiscal 2027. If Nvidia outperforms projections (as it has often done), its profits could grow far quicker, leading to a cheaper stock price. Its growth and size are hard to reconcile, which could make its future difficult to predict. But the company's China business is returning, and domestic AI demand is driving huge data center buildouts. These two factors could easily cause Nvidia to exceed expectations, making it a winning stock pick over the next few years. Should you buy stock in Nvidia right now? Before you buy stock in Nvidia, consider this: The Motley Fool Stock Advisor analyst team just identified what they believe are the for investors to buy now… and Nvidia wasn't one of them. The 10 stocks that made the cut could produce monster returns in the coming years. Consider when Netflix made this list on December 17, 2004... if you invested $1,000 at the time of our recommendation, you'd have $654,781!* Or when Nvidia made this list on April 15, 2005... if you invested $1,000 at the time of our recommendation, you'd have $1,076,588!* Now, it's worth noting Stock Advisor's total average return is 1,055% — a market-crushing outperformance compared to 183% for the S&P 500. Don't miss out on the latest top 10 list, available when you join Stock Advisor. See the 10 stocks » *Stock Advisor returns as of August 18, 2025 Keithen Drury has positions in Nvidia. The Motley Fool has positions in and recommends Nvidia. The Motley Fool has a disclosure policy. 3 Reasons I'm Buying Nvidia's Stock Like There's No Tomorrow was originally published by The Motley Fool


Bloomberg
an hour ago
- Bloomberg
Mike Pence Urges Trump to Rethink Nvidia Deal, Equity Stake in Intel
Former Vice President Mike Pence expressed concerns about the US government taking an equity stake in Intel Corp. and a cut of Nvidia Corp. 's AI chip sales to China. 'This was not a strategy that we employed during the Trump-Pence years,' Pence said during an interview Thursday on Bloomberg Television's Balance of Power. 'I have great concerns about having the US government take a position in, with gold shares in Nippon Steel or just the latest discussions about taking a percent of Intel.'