Cognitively Inspired Benchmark

Cognitively inspired benchmarks evaluate artificial intelligence models' abilities to perform tasks mirroring human cognitive development, such as understanding conservation of quantity, perspective-taking, and mechanical reasoning. Current research focuses on using these benchmarks to probe the capabilities of large vision-language models (LVLMs) and large language models (LLMs), often revealing discrepancies between model performance and human-like intelligence, particularly in higher-order cognitive functions. These benchmarks are crucial for identifying limitations in current AI architectures and guiding the development of more sophisticated models that better approximate human cognitive abilities, ultimately advancing both AI research and our understanding of human cognition itself.

Papers