The matching of manuscripts to journals is highly inefficient and imperfect. To start with, the editorial criteria of some journals may seem unclear to authors. Authors and editors (and even editors within the same team) might disagree on a manuscript’s suitability, or on the scientific advance of the work or its implications1. Also, editors may misjudge these considerations, or assess a manuscript on the basis of an ill-suited scientific context2. And authors may have an overly rosy view of their work, and are understandably incentivized to ‘aim high’ when choosing a journal.

All of these factors contribute to more submissions, rejections and re-submissions, to larger peer-reviewing efforts, to discouragement — particularly in younger academics — and to delayed dissemination of scientific outputs. However, when it comes to curating the literature, the system works3. But can it be made to operate more efficiently? Many scientists have long advocated for alternative systems, from removing the editor’s role as a curator to disentangling quality assessment from publishing3. Most attempts at changing scholarly publishing have remained so.

Generative artificial intelligence (AI) could move the needle. Can a suitably trained chatbot take on aspects of the role of an experienced journal editor? The chatbot would undoubtedly be tirelessly fast and stoical, and it would leverage more factual information than the world’s journal editors combined. However, it would be foolish for publishing houses to relinquish the entirety of editorial decisions to today’s generative AI systems: they can only provide plausible responses to prompts, and such plausibility cannot be guaranteed to follow truth nor logic. Still, these shortcomings do not imply uselessness, as shown in Box 1 by an example dialogue between an author and a publicly available large language model.

What can be inferred from this representative conversation? Firstly, the downsides: the chatbot’s ‘concerns’ are actually addressed in the report4,5 (which was included as part of the prompt to the chatbot) and earlier in the dialogue. However, a future multimodal system that can interpret scientific schematics, imagery, graphs and data may be less likely to confidently make erroneous assertions. Also, the questions asked by the chatbot are easily answerable from the text in the manuscript. Still, they are highly relevant to the quality of the work as would be judged by editors and reviewers. And some of the chatbot’s assertions — such as “the study design is rigorous” — are not to be taken at face value; the chatbot cannot assess actual rigour, it can only infer it from the manuscript’s text (and so would most readers, regardless of actual expertise).

However, the upsides of a future AI journal editor with enhanced skills are enticing. In particular, a chatbot fine-tuned with the journal’s historical output and editorial know-how, and reinforced with editorial feedback, could guide authors as to the degree of ‘fit’ of their work to the journal. It may help them craft a manuscript that more clearly highlights the most salient points. Moreover, it may make authors notice any shortcomings in the evidence or claims, or in the reporting of the methodology. Or the dialogue may make them realize that the manuscript would fare better in a more fitting journal.

At the same time, an AI journal editor might speed up editorial assessments. If authors approve such a chatbot’s summary of the manuscript and are satisfied with the chatbot’s questions and with the overall conversation, they are likely to agree to make it available to human editors, to facilitate their judgement of the work.

Generative AI is advancing toward levels of sophistication that make these considerations rather plausible. From a purely financial perspective, time spent in assessing rejected manuscripts before peer review is unproductive for journals — especially if they are highly selective. Although useful and specific feedback provided by dutiful editors to authors of manuscripts that are rejected is welcome by authors, from a journal-productivity viewpoint a more effective process than today’s workflow would be for editors to screen newly submitted manuscripts and to engage only with the authors of promising work; the authors of the unselected manuscripts would not receive a rejection message and would be free to take their manuscript elsewhere after a pre-specified number of days. But such a no-explicit-rejection process might be too big of a culture change for authors and editors; instead, a conversation with a chatbot that has been imbued with the journal’s editorial expertise would better conform to academic incentives and to expectations for feedback (if nothing else, confirming that an assessment process has been carried out).

Any practical implementation of an AI journal editor would involve lots of obstacles. Designing and implementing suitable ’guardrails’ and sanity checks would not be straightforward, and such an AI system could prove detrimental to breakthrough work that challenges current knowledge or practice. Also, AI journal editors might be easier to ‘game’ than most of the sentient sort. Moreover, AI editors could end up attracting substantially more unfruitful submissions. And it will escape no one’s attention that making editors more productive will reduce the number of them needed. Yet, will inspecting the performance of generative AI systems raise the need for counterfactual input and for auditors of the human kind?