On the importance of severely testing deep learning models of cognition

Jeffrey S. Bowers*, Gaurav Malhotra, Federico Adolfi, Marin Dujmović, Milton L. Montero, Valerio Biscione, Guillermo Puebla, John H. Hummel, Rachel F. Heaton

*Corresponding author for this work

Research output: Contribution to journalArticle (Academic Journal)peer-review

Abstract

Researchers studying the correspondences between Deep Neural Networks (DNNs) and humans often give little consideration to severe testing when drawing conclusions from empirical findings, and this is impeding progress in building better models of minds. We first detail what we mean by severe testing and highlight how this is especially important when working with opaque models with many free parameters that may solve a given task in multiple different ways. Second, we provide multiple examples of researchers making strong claims regarding DNN-human similarities without engaging in severe testing of their hypotheses. Third, we consider why severe testing is undervalued. We provide evidence that part of the fault lies with the review process. There is now a widespread appreciation in many areas of science that a bias for publishing positive results (among other practices) is leading to a credibility crisis, but there seems less awareness of the problem here.

Original languageEnglish
Article number101158
Number of pages11
JournalCognitive Systems Research
Volume82
Early online date22 Aug 2023
DOIs
Publication statusPublished - 1 Dec 2023

Bibliographical note

Publisher Copyright:
© 2023

Keywords

  • Memory
  • Neural networks
  • Perception
  • Psychology
  • Severe testing
  • Vision

Fingerprint

Dive into the research topics of 'On the importance of severely testing deep learning models of cognition'. Together they form a unique fingerprint.

Cite this