Research articles

Filter By:

Article Type
Year
  • With the rapid development of natural language processing (NLP) models in the last decade came the realization that high performance levels on test sets do not imply that a model robustly generalizes to a wide range of scenarios. Hupkes et al. review generalization approaches in the NLP literature and propose a taxonomy based on five axes to analyse such studies: motivation, type of generalization, type of data shift, the source of this data shift, and the locus of the shift within the modelling pipeline.

    • Dieuwke Hupkes
    • Mario Giulianelli
    • Zhijing Jin
    AnalysisOpen Access
  • The number of publications in artificial intelligence (AI) has been increasing exponentially and staying on top of progress in the field is a challenging task. Krenn and colleagues model the evolution of the growing AI literature as a semantic network and use it to benchmark several machine learning methods that can predict promising research directions in AI.

    • Mario Krenn
    • Lorenzo Buffoni
    • Michael Kopp
    AnalysisOpen Access
  • Training a deep neural network can be costly but training time is reduced when a pre-trained network can be adapted to different use cases. Ideally, only a small number of parameters needs to be changed in this process of fine-tuning, which can then be more easily distributed. In this Analysis, different methods of fine-tuning with only a small number of parameters are compared on a large set of natural language processing tasks.

    • Ning Ding
    • Yujia Qin
    • Maosong Sun
    AnalysisOpen Access