Cargando…
Human heuristics for AI-generated language are flawed
Human communication is increasingly intermixed with language generated by AI. Across chat, email, and social media, AI systems suggest words, complete sentences, or produce entire conversations. AI-generated language is often not identified as such but presented as language written by humans, raisin...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
National Academy of Sciences
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10089155/ https://www.ncbi.nlm.nih.gov/pubmed/36881628 http://dx.doi.org/10.1073/pnas.2208839120 |
_version_ | 1785022710197256192 |
---|---|
author | Jakesch, Maurice Hancock, Jeffrey T. Naaman, Mor |
author_facet | Jakesch, Maurice Hancock, Jeffrey T. Naaman, Mor |
author_sort | Jakesch, Maurice |
collection | PubMed |
description | Human communication is increasingly intermixed with language generated by AI. Across chat, email, and social media, AI systems suggest words, complete sentences, or produce entire conversations. AI-generated language is often not identified as such but presented as language written by humans, raising concerns about novel forms of deception and manipulation. Here, we study how humans discern whether verbal self-presentations, one of the most personal and consequential forms of language, were generated by AI. In six experiments, participants (N = 4,600) were unable to detect self-presentations generated by state-of-the-art AI language models in professional, hospitality, and dating contexts. A computational analysis of language features shows that human judgments of AI-generated language are hindered by intuitive but flawed heuristics such as associating first-person pronouns, use of contractions, or family topics with human-written language. We experimentally demonstrate that these heuristics make human judgment of AI-generated language predictable and manipulable, allowing AI systems to produce text perceived as “more human than human.” We discuss solutions, such as AI accents, to reduce the deceptive potential of language generated by AI, limiting the subversion of human intuition. |
format | Online Article Text |
id | pubmed-10089155 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | National Academy of Sciences |
record_format | MEDLINE/PubMed |
spelling | pubmed-100891552023-09-07 Human heuristics for AI-generated language are flawed Jakesch, Maurice Hancock, Jeffrey T. Naaman, Mor Proc Natl Acad Sci U S A Social Sciences Human communication is increasingly intermixed with language generated by AI. Across chat, email, and social media, AI systems suggest words, complete sentences, or produce entire conversations. AI-generated language is often not identified as such but presented as language written by humans, raising concerns about novel forms of deception and manipulation. Here, we study how humans discern whether verbal self-presentations, one of the most personal and consequential forms of language, were generated by AI. In six experiments, participants (N = 4,600) were unable to detect self-presentations generated by state-of-the-art AI language models in professional, hospitality, and dating contexts. A computational analysis of language features shows that human judgments of AI-generated language are hindered by intuitive but flawed heuristics such as associating first-person pronouns, use of contractions, or family topics with human-written language. We experimentally demonstrate that these heuristics make human judgment of AI-generated language predictable and manipulable, allowing AI systems to produce text perceived as “more human than human.” We discuss solutions, such as AI accents, to reduce the deceptive potential of language generated by AI, limiting the subversion of human intuition. National Academy of Sciences 2023-03-07 2023-03-14 /pmc/articles/PMC10089155/ /pubmed/36881628 http://dx.doi.org/10.1073/pnas.2208839120 Text en Copyright © 2023 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/This article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) . |
spellingShingle | Social Sciences Jakesch, Maurice Hancock, Jeffrey T. Naaman, Mor Human heuristics for AI-generated language are flawed |
title | Human heuristics for AI-generated language are flawed |
title_full | Human heuristics for AI-generated language are flawed |
title_fullStr | Human heuristics for AI-generated language are flawed |
title_full_unstemmed | Human heuristics for AI-generated language are flawed |
title_short | Human heuristics for AI-generated language are flawed |
title_sort | human heuristics for ai-generated language are flawed |
topic | Social Sciences |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10089155/ https://www.ncbi.nlm.nih.gov/pubmed/36881628 http://dx.doi.org/10.1073/pnas.2208839120 |
work_keys_str_mv | AT jakeschmaurice humanheuristicsforaigeneratedlanguageareflawed AT hancockjeffreyt humanheuristicsforaigeneratedlanguageareflawed AT naamanmor humanheuristicsforaigeneratedlanguageareflawed |