News & Comment

Filter By:

  • Most scientific prizes and medals are named after men, and most of these are also awarded to men. The very few awards named after women or not named after a person at all are more frequently awarded to women, although parity between the gender of recipients is still not achieved. We call on the scientific community to rethink the naming of academic awards, medals and prizes, their nomination and selection criteria, and to diversify awarding committees and procedures to ensure greater inclusivity.

    • Katja Gehmlich
    • Stefan Krause
    Comment
  • Large language models can be construed as ‘cognitive models’, scientific artefacts that help us to understand the human mind. If made openly accessible, they may provide a valuable model system for studying the emergence of language, reasoning and other uniquely human behaviours.

    • Michael C. Frank
    Comment
  • Large language models (LLMs) are impressive technological creations but they cannot replace all scientific theories of cognition. A science of cognition must focus on humans as embodied, social animals who are embedded in material, cultural and technological contexts.

    • Anthony Chemero
    Comment
  • Although artificial intelligence (AI) was already ubiquitous, the recent arrival of generative AI has ushered in a new era of possibilities as well as risks. This Focus explores the wide-ranging impacts of AI tools on science and society, examining both their potential and their pitfalls.

    Editorial
  • Large language models are capable of impressive feats, but the job of scientific review requires more than the statistics of published work can provide.

    • Grace W. Lindsay
    World View
  • In Japan, people express gratitude towards technology and this helps them to achieve balance. Yet, dominant narratives teach us that anthropomorphizing artificial intelligence (AI) is not healthy. Our attitudes towards AI should not be bult upon overarching universal models, argues Shoko Suzuki.

    • Shoko Suzuki
    World View
  • Algorithms are designed to learn user preferences by observing user behaviour. This causes algorithms to fail to reflect user preferences when psychological biases affect user decision making. For algorithms to enhance social welfare, algorithm design needs to be psychologically informed.

    • Carey K. Morewedge
    • Sendhil Mullainathan
    • Jens O. Ludwig
    Comment
  • The current debate surrounding the use and regulation of artificial intelligence (AI) in Brazil has social and political implications. We summarize these discussions, advocate for balance in the current debate around AI and fake news, and caution against preemptive AI regulation.

    • Cristina Godoy B. de Oliveira
    • Fabio G. Cozman
    • João Paulo C. Veiga
    Comment
  • Large language models (LLMs) do not distinguish between fact and fiction. They will return an answer to almost any prompt, yet factually incorrect responses are commonplace. To ensure our use of LLMs does not degrade science, we must use them as zero-shot translators: to convert accurate source material from one form to another.

    • Brent Mittelstadt
    • Sandra Wachter
    • Chris Russell
    Comment
  • State-of-the-art generative artificial intelligence (AI) can now match humans in creativity tests and is at the cusp of augmenting the creativity of every knowledge worker on Earth. We argue that enriching generative AI applications with insights from the psychological sciences may revolutionize our understanding of creativity and lead to increasing synergies in human–AI hybrid intelligent interfaces.

    • Janet Rafner
    • Roger E. Beaty
    • Jacob Sherson
    Comment
  • Generative artificial intelligence (AI) tools have made it easy to create realistic disinformation that is hard to detect by humans and may undermine public trust. Some approaches used for assessing the reliability of online information may no longer work in the AI age. We offer suggestions for how research can help to tackle the threats of AI-generated disinformation.

    • Stefan Feuerriegel
    • Renée DiResta
    • Nicolas Pröllochs
    Comment
  • If mistakes are made in clinical settings, patients suffer. Artificial intelligence (AI) generally — and large language models specifically — are increasingly used in health settings, but the way that physicians use AI tools in this high-stakes environment depends on how information is delivered. AI toolmakers have a responsibility to present information in a way that minimizes harm.

    • Marzyeh Ghassemi
    Comment