Scripta Materialia 193 (2020) 33–37 Contents lists available at ScienceDirect Scripta Materialia journal homepage: www.elsevier.com/locate/scriptamat Race against the Machine: can deep learning recognize microstructures as well as the trained human eye? Michiel Larmuseau a,b,c, , Michael Sluydts a,b , Koenraad Theuwissen d , Lode Duprez d , Tom Dhaene c , Stefaan Cottenier a,b a Center for Molecular Modeling, Ghent University, Technologiepark 46, Zwijnaarde, B-9052, Belgium b Department of Electromechanical, Ghent University, Technologiepark 46, Zwijnaarde, B-9052, Belgium c IDLab, Department of Information Technology, Ghent University – IMEC, Technologiepark 126, Zwijnaarde, B-9052, Belgium d OCAS NV/ArcelorMittal Global R&D Gent, Pres. J. F. Kennedylaan 3, Zelzate, B-9060, Belgium a r t i c l e i n f o Article history: Received 23 July 2020 Revised 8 October 2020 Accepted 12 October 2020 Keywords: Image analysis Steels Modeling Scanning electron microscopy (SEM) a b s t r a c t The promising results of deep learning in image recognition suggest a huge potential for microscopic analyses in materials science. One major challenge for its adoption in the study of materials is the lim- ited number of images that are available to train models on. Herein, we present a methodology to create accurate image recognition models with small datasets. By explicitly taking into account the magnifi- cation and by introducing appropriate transformations, we incorporate as many insights from material science in the model as possible. This allows for a highly data-efficient training of complex deep learning models. Our results indicate that a model trained with the presented methodology is able to outperform human experts. © 2020 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved. Microscopy images are an important source of information on the small-scale structure of materials, referred to as the mi- crostructure. However, most of the time, this information is anal- ysed only qualitatively: domain experts for instance mainly exam- ine these images to assess whether the processing of the material went well. While such an assessment is important, a lot of infor- mation contained in the microscopy image is not used. In recent years, approaches towards a more quantitative analysis of these images using machine learning have been thoroughly investigated in literature[1–4]. However, all these approaches rely on generic microstructure descriptors that are not tailored to the specific ma- terials in the dataset, resulting in sub-optimal performance. Deep learning methods[5] make it possible to learn microstruc- tural descriptors directly from the available data. Despite the first report of deep learning outperforming humans in an image recog- nition task already dating back to 2011[6], its adoption in practical material science remains limited[7–9]. A possible explanation for this, is that deep learning is often only deemed to outperform clas- sical methods when large datasets of images are available. With most commonly used datasets in material science literature con- taining around the order of a thousand images[2,4,10], this opinion feels reasonable. However, recent work[11] has shown that by cre- Corresponding author. E-mail address: michiel.larmuseau@ugent.be (M. Larmuseau). ating tailor-made deep learning models for specific datasets, one can outperform models that use generic microstructure descrip- tors. This would make it possible to investigate how well deep learn- ing models perform in recognizing microstructures compared to expert materials scientists. Although it is clear from other fields in computer vision that deep learning models can outperform human experts[12,13] provided a sufficient number of images, we here aim to investigate the performance of models that are trained on only around a hundred microstructure images in total. A hundred im- ages is a practical amount, as it can easily be collected in a sys- tematic study of a class of materials. To evaluate the performance of both a neural network that is obtained using the methodology presented in [11] and the panel of experts, we have organized two different quizzes. For the first quiz, 36 microscopy images need to be assigned to one of the five pre-defined classes of microstructures. The option “None of these” is included in case a microstructure image is shown that does not belong to one of the five pre-defined classes. The dataset on which the model was trained, the training set, contains 134 images in to- tal, with magnifications, expressed as pixels per micrometer, rang- ing from 1.1 to 212 pixels per micrometer (ppμm). Both optical mi- croscopy images and scanning electron microscopy (SEM) images have been included. An example for each type of microstructure is shown in Fig. 1 (a). For the second quiz, 21 SEM microscopy images of complex martensitic steels need to be assigned to one https://doi.org/10.1016/j.scriptamat.2020.10.026 1359-6462/© 2020 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.