AI and humans see objects differently: Meaning versus visual features

Representational embeddings inferred from human and DNN behavior. Credit: Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01041-7

While humans concentrate on the meaning of objects, artificial intelligence focuses on visual characteristics.

“These dimensions represent various properties of objects, ranging from purely visual aspects, like ’round’ or ‘white,’ to more semantic properties, like ‘animal-related’ or ‘fire-related,’ with many dimensions containing both visual and semantic elements,” explains Florian Mahner from the Max Planck Institute for Human Cognitive and Brain Sciences.

“Our results revealed an important difference: While humans primarily focus on dimensions related to meaning—what an object is and what we know about it—AI models rely more heavily on dimensions capturing visual properties, such as the object’s shape or color. We call this phenomenon ‘visual bias’ in AI.

“Even when AI appears to recognize objects just as humans do, it often uses fundamentally different strategies. This difference matters because it means that AI systems, despite behaving similarly to humans, might think and make decisions in entirely different ways, affecting how much we can trust them.”

Judgments for images

For human behavior, the scientists used about 5 million publicly available odd-one-out judgments over 1,854 different object images. For example, a participant would be shown an image of a guitar, an elephant, and a chair and would be asked which object doesn’t match. The scientists then treated multiple deep neural networks that can recognize images analogous to human participants and collected similarity judgments for images of the same objects used for humans.

Then, they applied the same algorithm to identify the key characteristics of these images—termed “dimensions” by the scientists—that underlie the odd-one-out decisions. By treating the neural network analogous to humans, this ensured direct comparability between the two.

“When we first looked at the dimensions we discovered in the deep neural networks, we thought that they actually looked very similar to those found in humans,” explains Martin Hebart, last author of the paper. “But when we started to look closer and compared them to humans, we noticed important differences.”

In addition to the visual bias identified by the scientists, they used interpretability techniques common in the analysis of neural networks for judging whether the dimensions they found actually made sense. For example, one dimension might feature a lot of animals and may be called “animal-related.”

To see if the dimension really responded to animals, the scientists ran multiple tests: They looked at what parts of the images were used by the neural network, they generated new images that best matched individual dimensions, and they even manipulated the images to remove certain dimensions. “All of these strict tests indicated very interpretable dimensions,” adds Mahner.

“But when we directly compared matching dimensions between humans and deep neural networks, we found that the network only really approximated these dimensions. For an animal-related dimension, many images of animals were not included, and likewise, many images were included that were not animals at all. This is something we would have missed with standard techniques.”

The scientists hope that future research will use similar approaches that directly compare humans with AI to better understand how AI makes sense of the world. “Our research provides a clear and interpretable method to study these differences, which helps us better understand how AI processes information compared to humans,” says Hebart. “This knowledge can not only help us improve AI technology but also provides valuable insights into human cognition.”

More information:
Florian P. Mahner et al, Dimensions underlying the representational alignment of deep neural networks with humans, Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01041-7

Provided by
Max Planck Society


Citation:
AI and humans see objects differently: Meaning versus visual features (2025, June 24)
retrieved 24 June 2025
from https://techxplore.com/news/2025-06-ai-humans-differently-visual-features.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.