Research articles

Filter By:

  • Deep learning is a powerful method to process large datasets, and shown to be useful in many scientific fields, but models are highly parameterized and there are often challenges in interpretation and generalization. David Gleich and colleagues develop a method rooted in computational topology, starting with a graph-based topological representation of the data, to help assess and diagnose predictions from deep learning and other complex prediction methods.

    • Meng Liu
    • Tamal K. Dey
    • David F. Gleich
    ArticleOpen Access
  • Continual learning is an innate ability in biological intelligence to accommodate real-world changes, but it remains challenging for artificial intelligence. Wang, Zhang and colleagues model key mechanisms of a biological learning system, in particular active forgetting and parallel modularity, to incorporate neuro-inspired adaptability to improve continual learning in artificial intelligence systems.

    • Liyuan Wang
    • Xingxing Zhang
    • Yi Zhong
    Article
  • Prediction of high-level visual representations in the human brain may benefit from multimodal sources in network training and the incorporation of complex datasets. Wang and colleagues show that language pretraining and a large, diverse dataset together build better models of higher-level visual cortex compared to earlier models.

    • Aria Y. Wang
    • Kendrick Kay
    • Leila Wehbe
    Article
  • Graph neural networks have proved useful in modelling proteins and their ligand interactions, but it is not clear whether the patterns they identify have biological relevance or whether interactions are merely memorized. Mastropietro et al. use a Shapley value-based method to identify important edges in protein interaction graphs, enabling explanatory analysis of the model mechanisms.

    • Andrea Mastropietro
    • Giuseppe Pasculli
    • Jürgen Bajorath
    Article
  • Halide perovskites are promising materials for light-emitting devices, given their narrowband emission and solution processability. However, detailed information on device degradation during operation is required to improve their stability, and this is challenging to obtain. Ji et al. propose a self-supervised deep learning method to capture multi-dimensional images of such devices in their operating regime faster than allowed by conventional imaging techniques.

    • Kangyu Ji
    • Weizhe Lin
    • Samuel D. Stranks
    ArticleOpen Access
  • The reconstruction of dynamic, spatial fields from sparse sensor data is an important challenge in various fields of science and technology. Santos et al. introduce the Senseiver, a deep learning framework that reconstructs spatial fields from few observations using attention layers to encode and decode sparse data, enabling efficient inference.

    • Javier E. Santos
    • Zachary R. Fox
    • Nicholas Lubbers
    ArticleOpen Access
  • Geometric deep learning has become a powerful tool in virtual drug design, but it is not always obvious when a model makes incorrect predictions. Luo and colleagues improve the accuracy of their deep learning model using uncertainty calibration and Bayesian optimization in an active learning cycle.

    • Yunan Luo
    • Yang Liu
    • Jian Peng
    Article
  • Human and animal motion planning works at various timescales to allow the completion of complex tasks. Inspired by this natural strategy, Yuan and colleagues present a hierarchical motion planning approach for robotics, using deep reinforcement learning and predictive proprioception.

    • Kai Yuan
    • Noor Sajid
    • Zhibin Li
    ArticleOpen Access
  • Organisms show complex behaviour resulting from a trade-off between obtaining information (explore) and using current information (exploit). Biswas et al. observe a mode-switching strategy modulated by sensory salience in a diverse range of organisms, including electric fish and humans, and argue that the observed heuristic could inform the design of active-sensing behaviours in robotics.

    • Debojyoti Biswas
    • Andrew Lamperski
    • Noah J. Cowan
    ArticleOpen Access
  • Prime editors are innovative genome-editing tools, but selecting guide RNAs with high efficiency remains challenging and requires costly experimental efforts. Liu and colleagues develop a method to design prime-editing guide RNAs based on transfer learning for in silico prediction of editing efficacy.

    • Feng Liu
    • Shuhong Huang
    • Wenjie Shu
    ArticleOpen Access
  • Learning causal relationships between variables in large datasets is an outstanding challenge in various scientific applications. Lagemann et al. introduce a deep neural network approach combining convolutional and graph models intended for causal learning in high-dimensional biomedical problems.

    • Kai Lagemann
    • Christian Lagemann
    • Sach Mukherjee
    ArticleOpen Access
  • Deep learning methods in natural language processing generally become more effective with larger datasets and bigger networks. But it is not evident whether the same is true for more specialized domains such as cheminformatics. Frey and colleagues provide empirical explorations of chemistry models and find that neural-scaling laws hold true even for the largest tested models and datasets.

    • Nathan C. Frey
    • Ryan Soklaski
    • Vijay Gadepally
    ArticleOpen Access
  • The immense amount of Wikipedia articles makes it challenging for volunteers to ensure that cited sources support the claim they are attached to. Petroni et al. use an information-retrieval model to assist Wikipedia users in improving verifiability.

    • Fabio Petroni
    • Samuel Broscheit
    • Sebastian Riedel
    ArticleOpen Access
  • With the rapid development of natural language processing (NLP) models in the last decade came the realization that high performance levels on test sets do not imply that a model robustly generalizes to a wide range of scenarios. Hupkes et al. review generalization approaches in the NLP literature and propose a taxonomy based on five axes to analyse such studies: motivation, type of generalization, type of data shift, the source of this data shift, and the locus of the shift within the modelling pipeline.

    • Dieuwke Hupkes
    • Mario Giulianelli
    • Zhijing Jin
    AnalysisOpen Access
  • The number of publications in artificial intelligence (AI) has been increasing exponentially and staying on top of progress in the field is a challenging task. Krenn and colleagues model the evolution of the growing AI literature as a semantic network and use it to benchmark several machine learning methods that can predict promising research directions in AI.

    • Mario Krenn
    • Lorenzo Buffoni
    • Michael Kopp
    AnalysisOpen Access