Abstract
Nowadays, the urgency for the highquality interdiffusion coefficients and atomic mobilities with quantified uncertainties in multicomponent/multiprincipal element alloys, which are indispensable for comprehensive understanding of the diffusioncontrolled processes during their preparation and service periods, is merging as a momentous trending in materials community. However, the traditional exploration approach for database development relies heavily on expertize and laborintensive computation, and is thus intractable for complex systems. In this paper, we augmented the HitDIC (highthroughput determination of interdiffusion coefficients, https://hitdic.com) software into a computation framework for automatic and efficient extraction of interdiffusion coefficients and development of atomic mobility database directly from large number of experimental composition profiles. Such an efficient framework proceeds in a workflow of automation concerning techniques of datacleaning, feature engineering, regularization, uncertainty quantification and parallelism, for sake of agilely establishing highquality kinetic database for target alloy. Demonstration of the developed infrastructures was finally conducted in fcc CoCrFeMnNi highentropy alloys with a dataset of 170 diffusion couples and 34,000 composition points for verifying their reliability and efficiency. Thorough investigation over the obtained kinetic descriptions indicated that the sluggish diffusion is merely unilateral interpretation over specific composition and temperature ranges affiliated to limited dataset. It is inferred that datamining over large number of experimental data with the combinatorial infrastructures are superior to reveal extremely complex composition and temperaturedependent thermal–physical properties.
Similar content being viewed by others
Introduction
Interdiffusion involves in a variety of materials processes in metallic solids, for instance, solidification^{1}, solid solution^{2}, aging^{3}, corrosion^{4}, mutual interaction between coatings and matrix^{5}, and so on. Diffusion represents the random motion of particles, i.e., atoms, molecules or ions, and is affected by the change of chemical potential of the systems, namely the interdiffusion or chemical diffusion. The movement of an individual particle is stochastic; however, it is governed by the thermodynamic and kinetic state of the system. In a complex system, the interactions between different types of particles are not identical. Such nonideal situation becomes serious as the number of components increases in systems with multiple components. For example, HEAs (highentropy alloys), where the multiple components are presented as principal constituents, nowadays serve as the alternatives for many traditional alloy systems, where only one or two components are presented as principal constituents. An interesting hotspot arises from the intriguing interactions, where diffusion rates seem to be rather low among the composition space around the equal atomic composition space. Such distinct diffusion behavior is reported to show the significant effect on the good mechanical properties and service performance of HEAs^{3}. Regarding the complex essence of diffusion, extensive efforts have been contributed to the investigation of the diffusion behaviors of HEAs for the very recent years^{6,7,8,9,10,11,12,13}. Debates about such sluggish diffusion effect still continue due to the incomplete overview of the desired systems^{14,15,16,17}. However, a comprehensive insight into diffusion behavior cannot be overwhelmingly supported without quantitative description of diffusion coefficients.
Determination of interdiffusion coefficients and later development of kinetic database have long been impeded by the incomplete techniques and toolsets. Matanobased methods are the most popular solutions for the determination of interdiffusion coefficient over the last several decades^{18}. Such historical methods and related tools are oriented for simple systems, i.e., binary^{19,20,21}, ternary^{22,23}, and a portion of higherorder systems^{24,25,26}; and thus, are inadequate to meet up with the interest of industrial research, where the systems are complex with multicomponents, e.g., solders, nickelbased alloys, HEAs and so on. Moreover, the size of the dataset is growing ever larger benefiting from advanced techniques for the preparation of diffusion couples^{27,28} and measurement of composition profiles^{29,30}. Assessment of diffusion description used to be based on 1–20 diffusion couples are shifting to hundreds, i.e., for Nibased alloys and HEAs. The amount of work towards datamining over such large dataset is challenging for the existing laborintensive procedures and workflow.
To date, the most promising approach for unveiling the complex kinetic interactions among multicomponent alloy systems has recently been described as the numerical inverse method^{18,31,32,33,34,35,36}. The essence of the numerical inverse method rests with revealing the inverse problem, that is reasoning the interdiffusion coefficients from the experimental composition profiles. The numerical inverse method dominates the Matanobased ones for proceeding reasoning without considering the number of components, but requiring the initial/final states and evolution rules^{18}. Currently, several efforts have been contributed to extract diffusion information from the specific diffusion dataset by some researchers independently^{31,32,33,34,35,36,37}. Unfortunately, the welldefined toolsets for numerical inverse methods are not yet enough to cope with dataset of large size for the complex systems. The primary issue is due to the naive implementation of numerical inverse method without revealing the deep essence of the inverse problem, where the tendency of being illposed is pressing. Situation gets worse as the size of dataset and the dimension of parameter space grows, turning into the largescale inverse problem^{38,39,40}. Curse of dimensionality haunts especially for multicomponent systems, because the number of descriptors for the related diffusion behaviors is large. Complying to CALPHAD (CALculation of PHAse Diagram) approach, the interaction parameters to be concerned for a quinary system can be up to 200 and even more. Assessments over such complex systems are difficult because the parameter space to be explored become myriad, while time expense for exploration become numerous. Such largescale inverse problem becomes intractable as it is much worse conditioned and often not uniquely solvable. When the parameter space and size of dataset reach a large scale, i.e., about 200 and more, pace would be extremely hard to proceed with manual construction based on expertize^{41}, which is neither agile for integral database development nor meeting up with features of highthroughput and automation^{42,43,44,45}.
Consequently, to bridge between the challenges and anxious anticipation, the infrastructures in automation fashion are thus in urgent need for the establishment of diffusion database, serving as the primary motivation of the present work. Highperformance computing (HPC) comes into the prior position to help with large dataset and speedup the related algorithms. Dimensionality reduction technique is required to simplify the overall complexity of the concerned diffusion descriptions. Meanwhile, both the uniqueness and generality should be emphasized towards the calculation results. Uncertainty quantification should also be served as an important portion, which indicates the useful information about the reliability of the assessed results. Integration of the proposed techniques are further in need of enabling a workflow of automation in practical applications and for accomplishing the thoughtful concerns above. Subsequently, we are going to report a successful demonstration of an automation computation framework for interdiffusivity evaluation and atomic mobility database development. It especially paves the way for settling the largescale inverse diffusion property problems with multicomponent and/or multiprincipal element alloys. Demonstration of the advanced infrastructure proposed in the present work is thus performed by conquering fcc CoCrFeMnNi HEAs from the point of view of its related diffusion behaviors.
Results and discussion
Framework and infrastructure
To be clear, the methods, strategies and codes are developed and bundled as HitDIC infrastructures, in the interest of realizing the interdiffusivity computation and atomic mobility dataset development in multicomponent/multiprincipal element alloys in a manner of automation. Originally, HitDIC is designed to extract interdiffusion coefficients from composition profiles and it has been successfully applied to multiple alloy systems^{10,12,46,47}. Later, HitDIC is featured with the capability of uncertainty quantification cooperating with Bayesian inference. However, the toolkits are not yet capable of dealing with dataset of large size and parameter space of high dimension. The wellestablished datamining techniques are, therefore, employed to levitate HitDIC to largescale inverse diffusion property problem. Infrastructure is built so as to drive the data/information flow in a manner of automation, as illustrated in Fig. 1a.
To begin with, dataset of composition profiles is extensively collected and preprocessed to produce denoised sample dataset. Handout validation can be proceeded by splitting sample sets into the training dataset and validation dataset with a ratio of 80–20. Secondly, the training dataset is further utilized in an optimization process, which offers the functionality for joint parameter selection and estimation. The optimization processes can be executed several times, and it will be repeated until a staged model with subset of effective parameters of interest is determined. Thirdly, the staged model will then be further proceeded to a regularization process, where the regularization term can be automatically tuned. Generality and stability of the estimations are intended to be improved in this period. Finally, in case that uncertainty of parameters of interest is concerned, the Bayesian inference will be employed to estimate the posterior distributions of the concerned parameters. Once the estimations of the concerned parameters and their uncertainties are determined, the diffusion database for the concerned system is thus developed. All the optimizers, i.e., the optimizer for joint parameter selection and estimation, regularization optimizer and the samplers for Markov chain Monte Carlo (MCMC), are supported by the cost evaluator with parallelism (see the “Methods”). Workflow of the concerned modules, i.e., joint parameter selection and parameter estimation, regularization and uncertainty quantification, are integrated. As the highperformance computing is in need, such kernels are implemented with C + + for the consideration of efficiency. The preprocessing and postprocessing are implemented with Python due to the plentiful toolkits for visualization and data manipulation, while the reports are generally presented in web view pages.
Data collection and preparation
To exemplify the capability of HitDIC infrastructure, HEAs are chosen as an example to demonstrate the capability of the developed infrastructure. The fcc CoCrFeMnNi alloys or Cantor alloys^{48} are employed in the present work. Being attracted by the sluggish diffusion effect, extensive investigations over the Cantor alloy, as well as other HEA systems, have been conducted by various groups^{6,7,8,9,10}. However, the diffusion database for the CoCrFeMnNi system is still considered incomplete. Therefore, Tsai et al.^{6} adopted the diffusion couple approach and a simplified Matanobased calculation method for studying diffusion coefficients on this system. Subsequent studies were carried out by Vaidya et al.^{9,49}, Kulkarni et al.^{50}, Verma et al.^{51}, Chen and Zhang^{11}, Wang et al.^{52}, Dąbrowa et al.^{8}, and Kucza et al.^{7} However, it is considered that those studies based on the limited dataset, i.e., <20 diffusion couples, are prone to be inconsistent with respect to the temperature or composition ranges. To produce trustworthy results, all reported data for the fcc CoCrFeMnNi system are therefore gathered, constituting a dataset of composition profiles with up to 170 groups of composition profiles over the temperature range of 1073–1373 K.
The collected composition profiles are further smoothed using the preset fitting functions, i.e., the logistic function or its superposition^{53}, the distribution functions and their superpositions^{54} and so on. Noises are thus removed, and the smoothed composition profiles are produced. By means of the fitting process, as detailed in Supplementary Methods, the dataset is renewed by applying the prior assumption that noise should be suspended. The composition profiles are thus resampled with the fitted functions, while the overall dataset possess up to 34,000 composition points, as illustrated in Fig. 2. It is indicated the concerned system is stable among composition range from 0.05 at. to 0.35 at. and temperature range from 1175 to 1373 K. It is important to note that the composition points remain sparse over the composition space in Fig. 2. Considering the large composition ranges, establishment of diffusion database for the concerned system is more desirable for completing the view of diffusion behaviors rather than outspreading the measured composition space with many more expensive diffusion couple/multiple experiments.
In this work, up to 34,000 composition points are employed, and the evaluated diffusion database is, therefore, considered generalizable among the mentioned composition ranges and temperatures ranges. What’s more, before taking the successive procedures, the overall dataset is then split into the training dataset and the validation dataset, with a ratio of 80 to 20.
Joint parameter selection and estimation
Joint parameter selection and parameter estimation are proceeded with the variableselection genetic algorithm. The default sets of atomic mobility parameters are assigned with the ones up to the first order for the binary systems and all those of ternary systems. For CoCrFeMnNi HEA, dimensionality of the concerned parameter space is large, e.g., roughly up to 200, which is about ten times more than that of binary or ternary systems. Redundant descriptors are likely to be introduced and desired to be sweep off due to convergency difficulty for optimizer towards problems of very high dimension. The variableselection genetic algorithm is subsequently applied to identify the most appropriate subset of parameters and their related estimations.
Convergence sequence of the proposed algorithm is illustrated in Fig. 3, where both Akaike information criterion (AIC) and Bayesian information criterion (BIC) are tested. In Fig. 3a, the AIC, residual summation square (RSS) and selection ratios are superimposed. The sequence of selection ratio converges nicely, reaching about 35% out of 200 parameters. The AIC drops rapidly at the early stage, however, results in a long tail as the algorithm proceeds. Although AIC sequence evolves slowly at the latter stage, the competition between parameters remains intensely, as indicated in Supplementary Fig. 2. With respect to the training dataset, the RSS and selection ratios evolve consistently with AIC value, indicating the effectiveness of the proposed joint parameter selection and estimation process. Furthermore, the sequence of RSS for the training and testing sets behaves similarly, which implies the validity of the variableselection genetic algorithm.
When it comes to the result based on BIC, more stringent selection efficiency is achieved, i.e., 18%, as shown in Fig. 3b. The potential reason for such distinct difference in selection ratios for the two criteria lies in that BIC imposes larger penalty on the number of concerned parameters. Comparing the optimized RSS for both criteria, the BIC succeeds to achieve by RSS_{BIC} = 1.67, which is slightly better than that of AIC, i.e., RSS_{AIC} = 1.71. That is when the BIC is employed, the result of variableselection genetic algorithm is able to achieve a better fitting goodness with less descriptors. It is indicated that the BIC is more suitable when the dimension is similar to the size of dataset. Currently, the result based on BIC is accepted as the product of the joint parameter selection and estimation and used in subsequent investigation. For details about the selected effective parameters and the obsoleted ones, the readers can refer to Supplementary Table 2. It is worthy of mentioning that the AIC might be not effective for the current size of the dataset; however, it might be effective for even larger dataset. Moreover, no matter what criterion is applied, the difficulty in proceeding the optimization remain intractable without welldesigned cost evaluator accommodated with highperformance parallel computing resources.
Regularization
In the framework of inverse problem, the estimated model has a fixed but unknown probabilistic relation to the data space. In previous researches on numerical inverse method, the solution to the inverse problem is found to be sensitive to the size of dataset, when the overall size of the dataset is small^{53,55}. Such phenomenon accords well with the primary feature of inverse problem, i.e., being illposed, resulting in severe problem about uniqueness. Plainly, the optimization using merely the first term in Eqs. (11) or (12) is insufficient to guarantee uniqueness of the solutions to inverse problem. The feature of being illposed can be weaken by increasing the sample sizes, i.e., considering as many as observations in the inverse process. As shown in Fig. 3, the solutions to the inverse problem can be reduced to limited alternatives, as they are constrained by means of expertize and statistic criteria as mentioned. Unfortunately, the potential solutions to inverse problem are still massive, inferred from the convergence sequences with long tail as presented in Fig. 3.
Therefore, a technique to address the problem of nonuniqueness is taken into consideration. Regularization is one of the common techniques served for releasing the illposedness of the inverse problem. The key of this technique is to introduce the concept of conditional wellposedness and shifts from searching for stable methods to reaching approximate solution with prior assumption. In other words, the regularization is to apply prior assumption on the solutions to the inverse problem, and therefore the target solution can be reduced to a limited model and the parameter space of less freedom.
The most frequently used prior assumption for regularization is that the L_{1} norm or L_{2} norm of the solution to inverse problem is considerably small enough. In practice, regularization is fulfilled by solving an optimization problem penalized by the L_{1} norm or L_{2} norm of the concerned parameters, where a regularization term, i.e., λ, is introduced to rescale the penalty. Presently, a workflow is used to tune the regularization term online while improving the solutions to the inverse problem (see the “Methods”). With the proposed algorithm, the selected estimators are further investigated, and the convergence sequences are presented in Fig. 4. From the point of view in regularization, the value of the prior assumptions, i.e., L_{1} norm or L_{2} norm, represents the complexity of the model or parameters. The key to regularization algorithm is to figure out an estimation of parameters with least complexity, while ensuring that the fitting goodness does not significantly turn worse.
For the L_{2} norm regularization, a significant increase of RSS value is observed when the regularization term grows larger than 5.40 × 10^{−7}, where the model complexity also decreases significantly comparing to the previous iterations, as indicated in Fig. 4b. Convergence sequence of the L_{1} norm regularization behave similarly as the L_{2} norm, and the turning point of the regularization term is the same, i.e., around 5.40 × 10^{−7}. Sequence of the RSS value for the validation dataset behaves consistently with that of training dataset, indicating the generality of estimations are good. Moreover, it also implies that the data distributions of the training set and validation set are similar, which generally exists when both datasets are sufficiently large. Such conclusion applies for both the L_{1} norm and L_{2} norm regularizations, as shown in Fig. 4a, b. As the regularizations are modeled based on the prior assumptions, both regularized results are considered reasonable and applicable for generalization and interpolation, as listed in the Supplementary Table 3.
Uncertainty quantification
Beyond the deterministic optimization procedures above, Bayesian inference is a suitable alternative for measuring the uncertainty of the solution to inverse problems. To further quantify the uncertainty of the concerned parameters, MCMC method, or more specifically Metropolis–Hastings algorithm, is currently used to infer their posterior means and variances (as demonstrated in methods). Considering acceptable computation cost, the multiple short but independent chains are employed currently. Twelve chains are proceeded with 50,000 iterations individually. Convergence diagnosis (see Supplementary Methods) is carried on the accepted sample points produced by the Metropolis–Hastings (MH) sampler. As shown in Fig. 5a, within sequence variances change significantly at the early period, while it achieves quite stable level as the sequences evolve. A better illustration of the stability of the Markov chains is indicated in Fig. 5b, where the between sequence variances are flat for different parameters. Unfortunately, the potential scale reduction factors for different parameters hardly reach 1, as illustrated in Fig. 5c, indicating that the individual chains are not yet reaching the stationary state. However, considering that the between sequence converge satisfactorily as shown in Fig. 5b, the current assessment is deemed as reasonable.
The proposed MH sampler totally draws 240,000 sample points with an accepted ratio of 40%. Posterior distributions of parameters are subsequently described as shown in Fig. 5d, where the mean and the related bounds of concerned parameters are imposed. It should be noted that the bounds are determined by a quantile of [0.2, 0.8], which covers about 60% sample points. Though only histograms of the posterior distributions of individual parameters are presented, they are actually subject to an integral joint posterior distribution. As shown in Fig. 5d, the posterior distributions of most parameters present in bell shapes, while the mean estimations reasonably rest around the highdensity region.
Estimated parameters with quantified uncertainties are important precursors to the quantification of uncertainties underpinning prediction and decisionmaking. For instance, the bounds of the predicted composition profiles can be retrieved as the uncertainty propagates through the forward problem of diffusion. As an example, Fig. 6a–d illustrates the fitted composition profiles of four selected diffusion couples of fcc CoCrFeMnNi HEAs, as denoted with dash lines and markers. The modelpredicted composition profiles are also imposed in Fig. 6, where the bounds of composition profiles rather than the exact optimal ones are presented. It has to be noted that the bounds are determined according to the parameters with uncertainties, which firstly propagate into the interdiffusion, and secondly to the composition profiles via the forward simulation. The overall goodness of the prediction to the experimental or fitted composition profiles are satisfactory, though parts of fitted composition profiles rest out of the bands of the predicted bounds, i.e., Fig. 6a. As the experimental procedures are taken as random events, the deviations outside the bounds infer the inconsistency with the assessed model. From the abroad comparison between the fitted and the modelpredicted composition profiles, the generality ability of the selected parameters and evaluated uncertainties are reasonable. For complete view of fitting goodness towards the whole dataset, the readers are referred to Supplementary Figs 10–31.
Remarks on optimization techniques
In the present work, the joint parameter selection and estimation have been succeeded to significantly reduce the dimensionality of the proposed problem. What’s more, the regularization is used to reduce the overall complexity of the selected model, while reserving the fitting goodness of the diffusion description. The two deterministic optimization strategies are feasible to come up with estimations, i.e., \({\mathbf{\theta }}_{{\mathrm{MLE}}}\), \({\mathbf{\theta }}_{{\mathrm{ReL}}1}\) and \({\mathbf{\theta }}_{{\mathrm{ReL}}2}\), with limited iterations. Besides, both techniques are constructed based on the genetic algorithm in the present work, which is famous for its robustness and promising ability for global optimal^{56}. The MCMC method is a statistic inference method, which is also capable of offering estimations, i.e., \({\mathbf{\theta }}_{{\mathrm{MAP}}}\), that produce the promising fitting goodness to the observations. However, it is generally more timeconsuming due to the considerable number of iterations. In most scenarios, the joint parameter selection and estimation followed with regularization are qualified, while the MCMC is superior to understand uncertainties of related model and parameters. Overall, the modelpredicted composition profiles to observations are similar for most samples or diffusion couples, as illustrated in Fig. 6e–h. For intuitive comparison between the observations and predictions, the readers are referred to the composition profiles presented in Supplementary Figs 32–53 for more detail.
However, it should be clarified that different optimization strategies actually perform differently according to their optimization criteria. Evidence lies in the statistics over the deviations between the observations and the predictions, i.e., as illustrated in Fig. 7. They are the distributions of the prediction biases due to variableselection genetic algorithm, L_{1}/L_{2} norm regularization and MCMC. The histograms of distributions are very similar to each other, though the estimations of the algorithms are quite different, as listed in Supplementary Table 3. Variableselection genetic algorithm with AIC and BIC, L_{1}/L_{2} norm regularization achieve similar mean square errors as denoted as star in Fig. 7. Among the three strategies, the variableselection genetic algorithm performs slightly better, as the prediction biases tend to concentrate more obvious around zero than the others. When it comes to the estimations produced by MCMC, the related fitting goodness is much better than the above three strategies as the related RSS is smaller. In trade of better overall fitting goodness, the prediction biases to the observations concentrate less significantly around 0 regarding the maximum aposterior.
Reason for difference in performance of different algorithms also lies in optimization criteria of the algorithms. Joint parameter selection and estimation is subjected to the information criteria, where the fitting goodness or RSS is partly concerned in the cost function. When most appropriate subset of parameters are selected, the estimated parameters are closed to the maximum likelihood estimation \({\mathbf{\theta }}_{{\mathrm{MLE}}}\). Regularization is expected to reduce the complexity of the concerned model by means of imposing additional penalty in the cost function. With the proposed regularization algorithm, results of regularization, i.e., \({\mathbf{\theta }}_{{\mathrm{ReL}}1}\) and \({\mathbf{\theta }}_{{\mathrm{ReL}}2}\), that do not significantly make worse prediction are taken. Therefore, the prediction performance of \({\mathbf{\theta }}_{{\mathrm{ReL}}1}\) and \({\mathbf{\theta }}_{{\mathrm{ReL}}2}\) is similar to that of \({\mathbf{\theta }}_{{\mathrm{MLE}}}\), however, the L_{1}/L_{2} norm shrinks distinctly. The goal of MCMC is to profile the joint posterior distribution of the concerned model and parameters. When the posterior distribution is reasonably drawn, the mean estimation is bound to the maximum aposterior \({\mathbf{\theta }}_{{\mathrm{MAP}}}\). The maximum aposterior tends to achieve better fitting goodness towards the entire dataset, though it behaves slightly worse towards the specific samples. The difference between \({\mathbf{\theta }}_{{\mathrm{MLE}}}\) and \({\mathbf{\theta }}_{{\mathrm{MAP}}}\) is partly due to the great convergence difficulty of the variableselection genetic algorithm in high dimension. Therefore, the maximum aposterior or the regularized estimations are more convincingly accountable estimations.
Nevertheless, MCMC starting from random initial proposals might take much longer iterations to reach the stationary state, when the initial estimations are far from the highdensity region. Situation gets worse when the dimension is extremely high and the dataset is especially large. Concerning the same problem, the expense of exploring the highdensity region for the deterministic algorithms is significant lower, which is superior to providing promising initial proposals for the MCMC algorithm. However, despite of all the pros and cons above, results of MCMC are indispensable as the information about the uncertainties can be provided. Analysis over the potential influence on the modelpredicted properties is therefore feasible. In more abroad applications, the uncertainty is favorable to territories of material design by offering decisionmaking insights^{57,58}.
Remarks on efficiency of parallelism
Among all the optimizers, the cost evaluator is fundamental to carry out predictions and evaluate the residual summation of square between the predictions and observations. For dataset of large size, evaluation of RSS is expensive, especially for optimizer demanded on great number of iterations, i.e., MCMC samplers. Parallelism is therefore mandatory while efficiency of parallelism is considerable. The scalability of a parallelism scheme is a measure of its ability to effectively exploit an increasing number of cores or processors on HPC clusters. Scalability analysis is usually designed for the most satisfactory algorithmarchitecture combination for a problem under different constraints on the growth of the problem size and the number of processors from the performance on fewer processors.
Efficiency test of the devised cost evaluator with parallelism is shown in Fig. 8b, where the case with 2 nodes (48 cores/threads on Intel Xeon CPU E52692) is set as the base. Comparison between the Intel Xeon CPU E52692 and AMD EPYC 7452 are considered, where the identical compilers, as well as the corresponding compiler options, are adopted. The cost evaluator scales nicely when a large dataset is considered for both computing resources, where 576 or more cores can be utilized. When the number of nodes is small, i.e., 4 nodes for E52692 and 1 node for EPYC 7452, the efficiency of the two types of computing resources are similar. However, the speedup of EPYC 7452 hits about 11 times with 9 nodes (576 processors) comparing to 8 times with 24 nodes (576 processors) with respect to E52692. Such a result benefits from a fact that, the machine with higher efficiency, i.e., high CPU clock cycle, works faster. Moreover, less message passing interface (MPI) communications between nodes would also benefit the efficiency for machines with more cores/threads on a single node. Currently, the optimization processes are carried on HPC clusters of AMD EPYC 7452.
Remarks on sluggish diffusion effect
For the HEAs, one of the most attracting topics is the existence of sluggish diffusion effect, which remains as a mystery in the past decade^{6,7,9,10,11,12,15,46,49}. Yeh^{59} originally proposed that kinetics of diffusion is hindered in comparison to pure metals and conventional alloys, resulting in smaller values of diffusion coefficients. It is inferred that the potential deduction of the diffusion rate of the highentropy alloy is due to the increase of entropy. To calibrate the influence of entropy on the diffusion rates, correlation between configurational entropy and different kinds of diffusion rates are examined. For sake of clarity, the thermodynamic description for the fcc CoCrFeMnNi system is considered identical to that of ideal solution phase, where only the configurational entropy rather than the excess interactions contributes to the thermodynamic factors.
To begin with, the effective tracer diffusion coefficients of pure metals or alloys of equal atomic compositions are compared. For pure metals or alloys of equal atomic compositions, larger configurational entropy is relatedly bound for higherorder system. Lacking in physical thermodynamic factors, the tracer diffusion coefficients are deemed as the effective ones, while the evaluated values using the assessed estimations \({\mathbf{\theta }}_{{\mathrm{MAP}}}\) are presented in Fig. 9a. The averages of the effective tracer diffusion coefficients of the quinary system are evaluated at various temperatures and taken as the base line for comparison, denoted as the dash lines in Fig. 9a. Considering the quaternary systems, the tracer diffusion coefficients fluctuate around the base line, which indicates that the reduction of entropy does not imply a firm tendency towards the acceleration of diffusion. Regarding the lower order systems, i.e., unary, binary, and ternary systems, the base plane rests around the middle among related tracer diffusion coefficients. It is thus concluded the entropy of configuration does not play a significant role in either hindering or accelerating the diffusion rates. In addition, the effect of averaging is rather obvious, as the deviation of the tracer diffusion coefficients shrinks as the number of components increases.
Owing to the absence of physical thermodynamic description, the effective tracer diffusion coefficients are further examined by comparing with the experimental results. The evaluated effective tracer diffusion coefficients are qualitatively compared to the ones measured by Tsai et al.^{6} and Vaidya et al.^{9,49}, as illustrated in Supplementary Figure 3. Among the tracer diffusion coefficients, \(D_{{\mathrm{Mn}}}^ \ast\) is dominant. The obtained effective tracer diffusion coefficient, i.e., \(D_{{\mathrm{Mn}}}^ \ast\), accords well with the results from Tsai et al.^{6} and Vaidya et al.^{9,49} However, results measured by Tsai et al. tends to underestimate \(D_{{\mathrm{Mn}}}^ \ast\) especially among the range of lower temperature. \(D_{{\mathrm{Cr}}}^ \ast\) measured by Tsai et al. and the present work are similar to each other, both of which are larger than the experimental results by Vaidya et al. As for \(D_{{\mathrm{Fe}}}^ \ast\), the results from Tsai et al. and Vaidya et al. agree well with each other, while the one measured in this work tends to be larger. Similar tendency applies to \(D_{{\mathrm{Ni}}}^ \ast\) with respect to \(D_{{\mathrm{Fe}}}^ \ast\). When it comes to \(D_{{\mathrm{Co}}}^ \ast\), Vaidya et al. come up with results that are smaller than the other two.
For the numerical inverse method, \(D_{{\mathrm{Fe}}}^ \ast\) and \(D_{{\mathrm{Cr}}}^ \ast\) are very similar, while \(D_{{\mathrm{Co}}}^ \ast\) and \(D_{{\mathrm{Ni}}}^ \ast\) are also similar. What’s more, difference among \(D_{{\mathrm{Fe}}}^ \ast\), \(D_{{\mathrm{Co}}}^ \ast\), \(D_{{\mathrm{Cr}}}^ \ast\) and \(D_{{\mathrm{Ni}}}^ \ast\) measured by Vaidya et al. are trivial, especially for the lower temperature range. The results measured by Tsai et al. for \(D_{{\mathrm{Fe}}}^ \ast\), \(D_{{\mathrm{Cr}}}^ \ast\), and \(D_{{\mathrm{Ni}}}^ \ast\) are also trivial with respect to each other. Overall, dependency of tracer diffusion coefficients on temperature are rather similar for all the components, indicating the similarity of their thermodynamic and kinetic behaviors. Unfortunately, the prior assumption of being the ideal solution is not fully applicable to the thermodynamic description of the fcc CoCrFeMnNi HEAs. From the point view of numerical inverse method, more profound thermodynamic description is, therefore, expected for the desired tracer diffusion coefficients from the research community, for sake of producing generalizable tracer diffusion coefficients of physical reliability among large composition ranges. Focusing on the correlation between diffusion rates and configuration entropy, the effective tracer diffusion coefficients are qualitatively reasonable notwithstanding its physical validity.
For more convincing evidence, the interdiffusion coefficients are adopted for characterizing the diffusion behaviors of fcc CoCrFeMnNi system and its related subsystems. Despite the number of components, different constituents for systems with the same components also contribute to the variation of the configurational entropy. To demonstrate the contribution of various constituents, the interdiffusion coefficients of fcc CoCrFeMnNi systems projected over various composition coordinates are evaluated.
Taking Fig. 10a, f, k and p as an example, main terms of interdiffusion coefficient matrices at 1073 K, i.e., \(\tilde D_{{\mathrm{CoCo}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{CrCr}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{FeFe}}}^{{\mathrm{Ni}}}\), and \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\), are projected over the composition coordinate of Co ranging from 0 to 0.2, according to the first column of subfigures in Fig. 10. When \(x_{{\mathrm{Co}}} = 0\), the matrix denotes fcc CrFeMnNi alloy; while \(x_{{\mathrm{Co}}} = 0.2\), the matrix denotes fcc CoCrFeMnNi alloy. The related entropy of ideal mixing is also imposed on the xaxis on the bottom, which can be evaluated according to
Among Fig. 10a, f, k and p, \(x_{\mathrm{E}}\) stands for the composition of component Co. With the increase of \(x_{{\mathrm{Co}}}\), the configurational entropy also increases. As shown, against the increment of entropy, \(\tilde D_{{\mathrm{CoCo}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{CrCr}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{FeFe}}}^{{\mathrm{Ni}}}\), and \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\) decreases significantly, implying a trend of being hindered by entropy.
When the projection falls to the composition coordinate of Cr, the response of the tracer diffusion coefficients to the variation of entropies are flat considering \(\tilde D_{{\mathrm{CrCr}}}^{{\mathrm{Ni}}}\), shown in Fig. 10g. However, \(\tilde D_{{\mathrm{CoCo}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{FeFe}}}^{{\mathrm{Ni}}}\), and \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\) drops as the related configurational entropy increases. As for the projection over the composition coordinate of Fe, \(\tilde D_{{\mathrm{CoCo}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{CrCr}}}^{{\mathrm{Ni}}}\), \(\tilde D_{{\mathrm{FeFe}}}^{{\mathrm{Ni}}}\), and \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\) shows limited variety against the related configurational entropy, though implying a tendency of reduction presented in Fig. 10c, h, m and r. From the point view of Mn composition coordinate, all interdiffusion coefficients are roused up as the configurational entropy presents a tendency of rising, though the magnitude of increment for \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\) is less significant, as illustrated in Fig. 10d, i, n and s. When it comes to the composition coordinate of Ni, the trending of \(\tilde D_{{\mathrm{CrCr}}}^{{\mathrm{Ni}}}\) and \(\tilde D_{{\mathrm{FeFe}}}^{{\mathrm{Ni}}}\) are rather flat with respect to the variation of entropy, referring to Fig. 10j and o. Moreover, controversial trends are observed for \(\tilde D_{{\mathrm{CoCo}}}^{{\mathrm{Ni}}}\) and \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\), as the former drops while the latter rises up.
The composition ranges of quaternary systems are covered by current training dataset, and thus, the interpolated interdiffusion coefficients for the quaternary systems are considered properly generalized. Considering correlation for temperatures other than 1073 K, similar conclusions can be drawn according to Supplementary Figs 4–6. As tendency of various interdiffusion coefficients against various projection coordinates remains similar to that of 1073 K. Regarding temperatures other than the specific ones, the related diffusion rates can be inferred from the Arrhenius relation of different components and systems, as listed in Supplementary Table 1. However, among the concerned temperature range, the interdiffusion coefficients do not perceive the tendency of sluggish diffusion. That is, the observations above do not launch a firm correlation between the variation of entropy and interdiffusion coefficients. Unfortunately, without a comprehensive overview of the diffusion rates, it is prone to unilaterally attribute the deduction of diffusion rates to the increment of entropy, i.e., \(\tilde D_{{\mathrm{CoCo}}}^{{\mathrm{Ni}}}\) and \(\tilde D_{{\mathrm{MnMn}}}^{{\mathrm{Ni}}}\) in Fig. 10e and t respectively.
A portion of the previous investigations over the diffusion behavior of HEAs reported that under the normalized temperature scale, the diffusion rates of the systems with higher entropy would be smaller^{6}. To further examine such a hypothesis, the effective tracer diffusion coefficients for various systems with equal atomic constituents at different normalized temperatures are evaluated for direct comparison. Assuming that alloys serve under 0.4T_{m} of the related systems, as shown in Fig. 9b, CoCrMnNi alloy is dominantly smaller than the rest, i.e., CoCrFeNi, CoFeMnNi, CrFeMnNi and CoCrFeMnNi alloys. Moreover, the effective tracer diffusion coefficients over the CoCrFeMnNi alloy rank beyond those of CoCrMnNi alloy, though the former is deemed as the one with higher entropy. Similar tendencies are found among the normalized temperatures, i.e., \(T/T_m = 0.6\) and \(T/T_m = 0.8\). Again, from the point of view in normalized temperature with respect to the melting point, the comparison result does not earn credit for the existence of sluggish diffusion.
Referring to averaging effect in Fig. 9a, the diffusion rates of various systems remain the same level of the fcc CoCrMnNi alloy. Among the concerned alloys, fcc CoCrMnNi alloy has the lowest melting temperature, i.e., 1500.82 K^{8}, while the melting temperatures for the others alloys are CoCrFeNi (1711 K^{8}), CoFeMnNi(1543 K^{8}), CrFeMnNi(1620 K^{60}) and CoCrFeMnNi(1572 K^{8}). It seems that the fcc CoCrMnNi system achieves the lowest diffusion rates at various normalized temperatures because of the lower melting point. That is, the normalized temperature, i.e., \(0.4T_{\mathrm{m}}^{{\mathrm{CoCrMnNi}}} = 600\,{\mathrm{K}}\), is significantly smaller than that of fcc CoCrFeMnNi alloy, i.e., \(0.4T_{\mathrm{m}}^{{\mathrm{CoCrFeMnNi}}} = 628\,{\mathrm{K}}\). As diffusion rate is subjected to the Arrhenius relation, it is not surprising that the alloy with lower melting point achieves lower diffusion rate with respect to the normalized temperature.
With the assessed diffusion descriptions of fcc CoCrFeMnNi system and its related subsystems, the concerned diffusion behaviors are able to be fully demonstrated by quantitative mathematical relations. As a conclusion due to rigor comparison, the sluggish diffusion of the fcc CoCrFeMnNi highentropy alloy remains no more than thermophysical state functions instead of mystery, which can be quantitatively evaluated with credible diffusion database found on large amount of experimental information.
Methods
Numerical inverse method
Concerning the diffusion processes, with adhoc thermokinetic description, the predictions to diffusion behaviors of mass can be revealed by solving diffusion equations. To fulfill such ambitions, the thermokinetic description is indispensable. The idea of tuning the most suitable kinetic description that accounting for observations, i.e., the experimental composition profiles, lies with the inverse problem, namely the numerical inverse method. The inverse problem of kinetic description can be generally casted into the framework of the partial differential equation constrained optimization problem,
where \({\mathbf{x}}_l\) denotes the experimental composition profile, while \({\tilde{\mathbf x}}_l\) the prediction to the composition profiles using the extended Fick’s second law. In Eq. (2), the prediction needs to be produced by means of solving the diffusion equations, i.e.,
for k ranging from 1 to n, where n is the number of concerned components in the system. Foremost, modeling of the interdiffusion coefficients is essential for construction of the forward problem and inverse problem. Currently, modeling of the interdiffusion coefficients is subject to the CALPHAD convention^{61}, as detailed in Supplementary Methods.
With the numerical inverse method, the interdiffusion coefficients of the concerned systems can be retrieved. Atomic mobility parameters are also available because the interdiffusion coefficients are parameterized following the CALPHAD convention. In the previous applications of numerical inverse method, both the interdiffusion coefficients and atomic mobility parameters are accessible, although the number of the diffusion couples involved in the calculation is less than 20. What’s more, when interdiffusion coefficients serves as the target of calculation, HitDIC performs nicely recovering the interdiffusion coefficients for the lower systems. Regarding to the growing number of diffusion couples in concerned dataset, advanced techniques and strategies are introduced in the present work for pursing diffusion database of highquality.
Variableselection genetic algorithm
We assume K as the competing parameters in the total parameter space \({\cal{M}}\) and a subset of them generates the observations, noted as x. Associated with all the parameters, there is a vector of parameters \({\mathbf{\theta }}\), i.e., \(\theta _1,\theta _2, \ldots ,\theta _K\). We can introduce an additional vector of selection parameters k, i.e., \(k_1,k_2, \ldots ,k_K\). The objective is to identify the true subset as well as to estimate the parameters associated with the subset,
where \(p\left( {{\mathbf{k}},{\mathbf{\theta }}{\mathbf{x}}} \right)\) is the posterior probability distribution. Each parameter of k is the indicator that takes the value 1 when the associated parameter comes in to force and is 0 otherwise. According to the Bayes’ rule, Eq. (2) is equivalent to Eq. (4), as
and
where \(p\left( {{\mathbf{x}}{\mathbf{k}},{\mathbf{\theta }}} \right)\) is the likelihood function. Here, \(\sigma\) is the variance of the residual between the predictions and observations. Generally, \(p\left( {{\mathbf{x}}{\mathrm{}}{\mathbf{\theta }},{\mathbf{k}}} \right)/p({\mathbf{\theta }},{\mathbf{k}})\) is taken as constant, though \(p\left( {{\mathbf{x}}{\mathrm{}}{\mathbf{\theta }},{\mathbf{k}}} \right)\) and \(p({\mathbf{\theta }},{\mathbf{k}})\) are not explicitly accessible.
In most genetic algorithms, only two main components are of problem dependence^{62}, i.e., problem encoding and evaluation function. In order to accommodate the problem of parameter selection and parameter estimation for automatic evaluation of interdiffusion coefficients and atomic mobilities, the binary encoding is adopted in the present work,
where H is the proposed encoding function. The proposed encoding pattern is schematically illustrated in Fig. 1b, where the selection parameters are laid out after the model parameters. Each selection parameter takes only one allele, while the number of bits occupied by individual model parameter may vary according to desired precision. The laidout of the encoding string has potential influence on the effects of the operators of the genetic algorithm^{63}, i.e., mutation and crossover. The crossover operator includes different schemata, i.e., singlepoint, twopoint, uniform and arithmetic crossover, while the singlepoint/twopoint crossover operator is adopted in the present work to retain the robustness of the selection and optimization processes. To endow sufficient ability for evolution, the mutation is uniformly considered for all bits via bit inversion with a ratio of mutation about 1%.
The canonical genetic algorithm is responsible for parameter estimation regarding to the evaluation and fitness of the problem. To fulfill the target of parameter selection, the evaluation or the objective function must be surrogated by introducing penalty on the number of effective model parameters. As one of the fancy evolutionary algorithms, the genetic algorithm possesses the features of being scalable and being flexible to consider many criteria in the optimization processes. Referring to solution with exhaustive selection and scoring scheme, the potential criteria, i.e., Ftest, information criteria and regularization, are the potential options. In general, the fitness function is used for genetic algorithm for measuring the driving force for evolution. Fitness is different from the objective functions, noted as \({\mathrm{OBJ}}({\mathbf{\theta }})\). The RSS, i.e., Eq. (2), is one of most popular options for the objective function, where the least RSS is generally pursued. Fitness function represents the probability of survival for the population of the solutions, therefore, larger fitness values are more desirable for the selection operators. Proper conversion is thus in need between the objective functions and the fitness function. For a population with P individuals, the fitness function, F, can be defined as
With the fitness function, the selection operator for genetic algorithm can therefore be conducted using the roulette over the fitness sequence.
One of the simplest and most convenient objective functions regarding the parameter selection is the information criterion, which concerns the model complexity and fitting goodness simultaneous, i.e., the AIC or BIC,
or
where K is the number of the effective parameters, N is the number of observations and \(\hat L\) is maximum value of the likelihood function. Generally, \(\hat L\) cannot be directly accessible, though it can be related to RSS as detailed in Supplementary Methods. Meanwhile, a study case for benchmarking is available in Supplementary Discussion.
Regularization optimizer with automatic hyperparameter tuning
The regularization is generally served as a powerful tool for preventing the overfitting while improving the generality of the assessed model and also estimated parameters^{64,65}. It is an important concept in the inverse problem, machine learning and so on. The most common strategy for regularization is to construct a surrogated objective function by introducing penalty on L_{1}/ L_{2} norm to the original objective function, i.e., Eq. (2). With a regularization term λ, the objective function can be reformulated as
for L_{1} regularization, and as
for L_{2} regularization. By determining an appropriate scale for the regularization term, the results that balance the extrapolation and explanatory capability of the proposed parameters/models are to be acquired, when the arguments of the minimum of the surrogated objective functions are resolved.
However, the regularization term is a tricky hyperparameter, which deserves meticulous tuning^{64,66,67}. An algorithm for automatically tuning the regularization term is proposed, as shown in Fig. 1c. Firstly, the estimated effective parameters and their estimations are imported from the variableselection genetic algorithm. The least RSS ported from variableselection GA is then used to estimate initial regularization term. The workflow then proceeds into a subroutine where the most appropriate estimations, i.e., \({\mathbf{\theta }}_{{\mathrm{ReL}}1}\) or \({\mathbf{\theta }}_{{\mathrm{ReL}}2}\), are pursued until no significant change takes place between subsequent iterations. The RSS value will be doublechecked to verify whether there is a significant increase in RSS of the training dataset. The workflow will be terminated once the current RSS value surpasses the least RSS at a certain degree. Or the regularization term will be increased, and the subroutine to determine new alternative estimations would be repeated. During the iterations, the least RSS will be updated as if a new alternative with smaller RSS occurs.
Metropolis–Hastings sampler with multiple independent chains
For the nonlinear inverse problem, the Bayesian inference might be the only tool available for quantifying the uncertainty of the concerned model and parameters. MCMC is one of the important interference tools based on the Bayes’ rule^{68},
The posterior distribution, i.e., \(p\left( {{\mathbf{\theta }}{\mathbf{x}}} \right)\), is generally not able to produce estimations, i.e., \({\hat{\mathbf \theta }}\), directly, however, it can be employed in a reversible Markov process during the Monte Carlo simulation. The samples are then drawn from the target distribution, while the posterior estimations can, therefore, be obtained. However, the Bayesian inference with the MCMC method is generally challenging for model with large parameter space and dataset of large sample size, considering the cost of time and computing expenses^{69}. This is partly due to the computational cost of such methods, since the evaluation of the objective function, i.e., Eq. (5) or Eq. (13), is generally computing expensive. Sufficient random walks are expected such that the obtained posterior distribution reaches a stationary state, which might require tens or hundreds of times of iterations more than the optimization or regularization processes.
In the naive Metropolis–Hastings algorithm, the posterior distribution is evaluated with respect to all the samples in the dataset, assuming that all samples, i.e., x, are independently measured. In addition, the multiple independent chains, i.e., M, are employed in the present work to draw samples from the posterior distribution of m parameters. One very long chain for MCMC is not applicable for a tall dataset due to time efficiency, and thus increasing the number of chains and running the chains in parallelism would be a promising alternative. The overall workflow of the developed sampling kernel is illustrated in Fig. 1d. It is worthy of mentioning that the initial estimations for the MCMC sequences are ported from those produced by regularization processes in order to let the MCMC chains locate around the high probability regions.
Implicit solver for multicomponent diffusion equations
In the inverse problem, the solver to the forward problem is extremely essential for the inverse process. To ensure the stability of the forward simulation process, a fully implicit finite difference scheme is applied to relax the stable condition constrained by step sizes of space and time. Prior to the demonstration of the proposed scheme, benchmarks are available in Supplementary Methods.
On a onedimension domain representing a diffusion couple, taking the ith grid node as an example, the conjugated grids are \(l_{i  1}\) and \(l_{i + 1}\), and \(h_i^ \) and \(h_i^ +\) are the spacings before and behind the current grid node. Imposing finite difference scheme on Eq. (3), the recursive formula for the pth component of a system with M components can be formulated
where \(\Delta t\) is the time step. To fulfill numerical simulation, the coefficient terms can be rewritten in the form of matrix and vector,
where \({\mathbf{A}}_{pp}^{t + 1}\) is the coefficient matrix on the lefthand side of Eq.(14), \({\mathbf{b}}_p^{t + 1}\) is the righthand side of Eq.(14). For zeroflux boundary condition,
and
Equations (14)–(20) subject to a set of linear equations, where the coefficient matrices are in the form of tridiagonal matrices, i.e.,
which can be solved easily with tridiagonal matrix algorithm. In case that \({\mathbf{A}}_{pp}^{t + 1}\) and \({\mathbf{b}}_p^{t + 1}\) are implicit, additional operations are required to estimate the two terms for each time step as illustrated in the Algorithm 1 of Supplementary Methods.
Parallelism of cost evaluator
The bottleneck of the efficiency for the numerical inverse method lies in the timeconsuming process of evaluating the objective function. In the framework of HitDIC software, a cost evaluator is responsible for calling the solver to diffusion equations to produce predictions, and calculating the deviation between the predictions and observations for the output of the objective function. For the sample dataset of large size, the evaluation of objective function is computationally expensive, which may be unfeasible without HPC. To assess the computing resources on HPC, the parallelism with MPI technique is adopted for the cost evaluator, as illustrated in Fig. 8a. The proposed parallelism mechanism behind the cost evaluator is suitable for genetic algorithm, MCMC with multiple chains and so on. Taking the genetic algorithm as an example, each iteration is required to evaluate cost for multiple guesses in a population, while the cost evaluator for each guess relies on the predictions to multiple samples in the training dataset. More specifically, the guesses and the samples are firstly rearranged to form the sequence of tasks in the master node. Once tasks are ready, the signal from the master node will be sent to the workers in the worker pool, while the workers will then offload tasks from master node repeatedly. In an active worker process, the cost evaluator would be executed in sequence, while the value of objective function will then be transferred back to the master. State of the worker will then be flushed and it will wait in the pool for the remaining tasks. The master node is responsible for offloading tasks and collecting the results to/from the computing nodes, and returning the results to different optimization solvers.
Data availability
The key data that support the findings within this paper can be found at the GitHub address https://github.com/zhongjingjogy/fcc_CoCrFeMnNi, and other data are available from the corresponding author upon reasonable request.
Code availability
Related algorithms are bundled in the HitDIC software, the latest version of which can be accessible from https://hitdic.com.
References
Takaki, T. et al. Primary arm array during directional solidification of a singlecrystal binary alloy: largescale phasefield study. Acta Mater. 118, 230–243 (2016).
Reed, R. C. The Superalloys: Fundamentals and Applications. (Cambridge University Press, 2006).
Yeh, J.W. et al. Nanostructured highentropy alloys with multiple principal elements: novel alloy design concepts and outcomes. Adv. Eng. Mater. 6, 299–303 (2004).
Ta, N., Zhang, L., Li, Q., Lu, Z. & Lin, Y. Hightemperature oxidation of pure Al: kinetic modeling supported by experimental characterization. Corros. Sci. 139, 355–369 (2018).
Clarke, D. R., Oechsner, M. & Padture, N. P. Thermalbarrier coatings for more efficient gasturbine engines. MRS Bull. 37, 891–898 (2012).
Tsai, K.Y., Tsai, M.H. & Yeh, J.W. Sluggish diffusion in Co–Cr–Fe–Mn–Ni highentropy alloys. Acta Mater. 61, 4887–4897 (2013).
Kucza, W. et al. Studies of “sluggish diffusion” effect in CoCrFeMnNi, CoCrFeNi and CoFeMnNi high entropy alloys; determination of tracer diffusivities by combinatorial approach. J. Alloy. Compd. 731, 920–928 (2018).
Dąbrowa, J. et al. Demystifying the sluggish diffusion effect in high entropy alloys. J. Alloy. Compd. 783, 193–207 (2019).
Vaidya, M., Pradeep, K. G., Murty, B. S., Wilde, G. & Divinski, S. V. Bulk tracer diffusion in CoCrFeNi and CoCrFeMnNi high entropy alloys. Acta Mater. 146, 211–224 (2018).
Chen, S., Li, Q., Zhong, J., Xing, F. & Zhang, L. On diffusion behaviors in face centered cubic phase of AlCoCrFeNiTi highentropy superalloys. J. Alloy. Compd. 791, 255–264 (2019).
Chen, W. & Zhang, L. Highthroughput determination of interdiffusion coefficients for CoCrFeMnNi highentropy alloys. J. Phase Equilib. Diffus. 38, 457–465 (2017).
Wang, R., Chen, W., Zhong, J. & Zhang, L. Experimental and numerical studies on the sluggish diffusion in face centered cubic CoCrCuFeNi highentropy alloys. J. Mater. Sci. Technol. 34, 1791–1798 (2018).
Choi, W.M., Jo, Y. H., Sohn, S. S., Lee, S. & Lee, B.J. Understanding the physical metallurgy of the CoCrFeMnNi highentropy alloy: an atomistic simulation study. NPJ Comput. Mater. 4, 1 (2018).
Dąbrowa, J. & Danielewski, M. Stateoftheart diffusion studies in the high entropy alloys. Metals 10, 347 (2020).
Divinski, S. V., Pokoev, A. V., Esakkiraja, N. & Paul, A. A mystery of “sluggish diffusion” in highentropy alloys: the truth or a myth? Diffus. Found. 17, 69–104 (2018).
Zhang, C. et al. Understanding of the elemental diffusion behavior in concentrated solid solution alloys. J. Phase Equilib. Diffus. 38, 434–444 (2017).
Beke, D. & Erdélyi, G. On the diffusion in highentropy alloys. Mater. Lett. 164, 111–113 (2016).
Zhong, J., Chen, L. & Zhang, L. Highthroughput determination of highquality interdiffusion coefficients in metallic solids: a review. J. Mater. Sci. 55, 10303–10338 (2020).
Matano, C. On the relation between the diffusioncoefficients and concentrations of solid metals. Jpn. J. Appl. Phys. 8, 109–113 (1933).
Wagner, C. The evaluation of data obtained with diffusion couples of binary singlephase and multiphase systems. Acta Metall. 17, 99–107 (1969).
Sauer, F. & Freise, V. Diffusion in binären Gemischen mit Volumenänderung. Z. für. Elektrochemie, Ber. der Bunsenges. für. Physikalische Chem. 66, 353–362 (1962).
Kirkaldy, J. S. & Young, D. J. Diffusion in the Condensed State. (Institute of Metals, London, 1987).
Whittle, D. & Green, A. The measurement of diffusion coefficients in ternary systems. Scr. Metall. 8, 883–884 (1974).
Paul, A. A pseudobinary approach to study interdiffusion and the Kirkendall effect in multicomponent systems. Philos. Mag. 93, 2297–2315 (2013).
Esakkiraja, N. & Paul, A. A novel concept of pseudo ternary diffusion couple for the estimation of diffusion coefficients in multicomponent systems. Scr. Mater. 147, 79–82 (2018).
Esakkiraja, N., Pandey, K., Dash, A. & Paul, A. Pseudobinary and pseudoternary diffusion couple methods for estimation of the diffusion coefficients in multicomponent systems and high entropy alloys. Philos. Mag. 99, 2236–2264 (2019).
Zhao, J.C., Zheng, X. & Cahill, D. G. Highthroughput diffusion multiples. Mater. Today 8, 28–37 (2005).
Xu, H. et al. Determination of accurate interdiffusion coefficients in fcc AgIn and AgCuIn alloys: a comparative study on the Matano method with distribution function and the numerical inverse method with HitDIC. J. Alloy. Compd. 798, 26–34 (2019).
Kodentsov, A. A., Bastin, G. F. & van Loo, F. J. J. in Methods for Phase Diagram Determination 222–245 (Elsevier, 2007).
Kodentsov, A. & Paul, A. in Handbook of Solid State Diffusion, Vol 2 207–275 (Elsevier, 2017).
Chen, W., Zhang, L., Du, Y., Tang, C. & Huang, B. A pragmatic method to determine the compositiondependent interdiffusivities in ternary systems by using a single diffusion couple. Scr. Mater. 90–91, 53–56 (2014).
Chen, W., Zhong, J. & Zhang, L. An augmented numerical inverse method for determining the compositiondependent interdiffusivities in alloy systems by using a single diffusion couple. MRS Commun. 6, 295–300 (2016).
Kucza, W. A combinatorial approach for extracting thermokinetic parameters from diffusion profiles. Scr. Mater. 66, 151–154 (2012).
Bouchet, R. & Mevrel, R. A numerical inverse method for calculating the interdiffusion coefficients along a diffusion path in ternary systems. Acta Mater. 50, 4887–4900 (2002).
Chen, Z., Zhang, Q. & Zhao, J.C. pydiffusion: A Python library for diffusion simulation and data analysis. J. Open Res. Softw. 7, 13 (2019).
Gaertner, D. et al. Concentrationdependent atomic mobilities in FCC CoCrFeMnNi highentropy alloys. Acta Mater. 166, 357–370 (2019).
Zhang, Q. & Zhao, J.C. Extracting interdiffusion coefficients from binary diffusion couples using traditional methods and a forwardsimulation method. Intermetallics 34, 132–141 (2013).
Biegler, L. et al., eds. LargeScale Inverse Problems and Quantification of Uncertainty. (John Wiley & Sons, 2011).
Chung, J., Knepper, S. & Nagy, J. G. in Handbook of Mathematical Methods in Imaging 47–90 (Springer New York, 2015).
Cullen, M., Freitag, M. A., Kindermann, S., & Scheichl, R. eds. Large Scale Inverse Problems: Computational Methods and Applications in the Earth Sciences. (De Gruyter, 2013).
Zhang, L. & Chen, Q. in Handbook of Solid State Diffusion, Vol.1. 321–362 (Elsevier, 2017).
Olson, G. B. & Kuehmann, C. J. Materials genomics: from CALPHAD to flight. Scr. Mater. 70, 25–30 (2014).
National Research Council, Division on Engineering and Physical Sciences, National Materials Advisory Board & Committee on Integrated Computational Materials Engineering. Integrated Computational Materials Engineering: A Transformational Discipline for Improved Competitiveness and National Security. (National Academies Press, 2008).
Nikolaev, P. et al. Autonomy in materials research: a case study in carbon nanotube growth. NPJ Comput. Mater. 2, 1–6 (2016).
Ozaki, Y. et al. Automated crystal structure analysis based on blackbox optimisation. NPJ Comput. Mater. 6, 75 (2020).
Li, Q. et al. On sluggish diffusion in Fcc Al–Co–Cr–Fe–Ni highentropy alloys: an experimental and numerical study. Metals 8, 16 (2017).
Chen, J. & Zhang, L. Compositiondependent interdiffusivity matrices in face centered cubic Ni–Al–X (X = Rh and W) alloys at 1423, 1473 and 1523 K: A highthroughput experimental measurement. Calphad 60, 106–115 (2018).
Cantor, B., Chang, I. T. H., Knight, P. & Vincent, A. J. B. Microstructural development in equiatomic multicomponent alloys. Mater. Sci. Eng. A 375–377, 213–218 (2004).
Vaidya, M., Trubel, S., Murty, B. S., Wilde, G. & Divinski, S. V. Ni tracer diffusion in CoCrFeNi and CoCrFeMnNi high entropy alloys. J. Alloy. Compd. 688, 994–1001 (2016).
Kulkarni, K. & Chauhan, G. P. S. Investigations of quaternary interdiffusion in a constituent system of high entropy alloys. AIP Adv. 5, 097162 (2015).
Verma, V., Tripathi, A. & Kulkarni, K. N. On interdiffusion in FeNiCoCrMn high entropy alloy. J. Phase Equilib. Diffus. 38, 445–456 (2017).
Wang, R. On the Determination of Diffusion Coefficients and Sluggish Diffusion Effect of Facecentered Cubic CoCrFeNiX(X=Mn,Cu) High Entropy Alloys. (Central South University, 2018).
Zhong, J., Chen, W. & Zhang, L. HitDIC: a freeaccessible code for highthroughput determination of interdiffusion coefficients in single solution phase. Calphad 60, 177–190 (2018).
Wei, M. & Zhang, L. Application of distribution functions in accurate determination of interdiffusion coefficients. Sci. Rep. 8, 5071 (2018).
Zhong, J., Zhang, L., Wu, X., Chen, L. & Deng, C. A novel computational framework for establishment of atomic mobility database directly from composition profiles and its uncertainty quantification. J. Mater. Sci. Technol. 48, 163–174 (2020).
McCall, J. Genetic algorithms for modelling and optimisation. J. Comput. Appl. Math. 184, 205–222 (2005).
Kochenderfer, M. J. Decision Making Under Uncertainty: Theory and Application. (MIT Press, 2015).
Lookman, T., Balachandran, P. V., Xue, D. & Yuan, R. Active learning in materials science with emphasis on adaptive sampling using uncertainties for targeted design. NPJ Comput. Mater. 5, 21 (2019).
Yeh, J.W. Recent progress in high entropy alloys. Ann. Chim. Sci. Mat. 31, 633–648 (2006).
Lederer, Y., Toher, C., Vecchio, K. S. & Curtarolo, S. The search for high entropy alloys: a highthroughput abinitio approach. Acta Mater. 159, 364–383 (2018).
Andersson, J. & Ågren, J. Models for numerical treatment of multicomponent diffusion in simple phases. J. Appl. Phys. 72, 1350–1355 (1992).
Whitley, D. A genetic algorithm tutorial. Stat. Comput. 4, 65–85 (1994).
Bhandari, D., Murthy, C. & Pal, S. K. Genetic algorithm with elitist model and its convergence. Int. J. Pattern Recogn. 10, 731–747 (1996).
Poggio, T., Torre, V. & Koch, C. Computational vision and regularization theory. Nature 317, 314–319 (1985).
Girosi, F., Jones, M. & Poggio, T. Regularization theory and neural networks architectures. Neural Comput. 7, 219–269 (1995).
Hansen, P. C. & O’Leary, D. P. The use of the Lcurve in the regularization of discrete illposed problems. SIAM J. Sci. Comput. 14, 1487–1503 (1993).
Engl, H. W., Hanke, M. & Neubauer, A. Regularization of Inverse Problems. Vol. 375 (Springer Science & Business Media, 1996).
Hewson, P. Statistical rethinking: a Bayesian course with examples in R and Stan. J. R. Stat. Soc. A Stat. 179, 1131 (2016).
Robert, C. P., Elvira, V., Tawn, N. & Wu, C. Accelerating MCMC algorithms. WIREs Comput. Stat. 10, e1435 (2018).
Acknowledgements
The financial support from the National MCF Energy R&D Program of China (Grant No. 2018YFE0306100), the Guangdong Province KeyArea Research and Development Program of China (2019B010943001), the Youth Talent Project of Innovationdriven Plan at Central South University (Grant No. 2282019SYLB026), and the Hunan Provincial Science and Technology Program of China (Grant No. 2017RS3002)Huxiang Youth Talent Plan is acknowledged. Jing Zhong acknowledges the support from the Fundamental Research Funds for the Central Universities of Central South University (Grant No. 2018zzts129). This is also part of Dr. Jing Zhong’s postdoctoral research work at Central South University, China. Jing Zhong acknowledges Dr. Richard Otis for discussing the strategy for parameter selection during the 47^{th} CALPHAD conference in Mexico.
Author information
Authors and Affiliations
Contributions
L.Z. conceived the presented idea and provided necessary materials. J.Z. designed, developed, and maintained the HitDIC infrastructure, data management system. J.Z. and L.Z. wrote the manuscript. All the authors, J.Z., L.C., and L.Z., discussed the results and commented on the manuscript.
Corresponding author
Ethics declarations
Competing interests
HitDIC is a freeaccessible software independently developed by current authors. The proposed algorithms and strategies are bundled with HitDIC, while the related results are computed with HitDIC and its related infrastructures. Therefore, the authors have no more competing interests to clarify.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Zhong, J., Chen, L. & Zhang, L. Automation of diffusion database development in multicomponent alloys from large number of experimental composition profiles. npj Comput Mater 7, 35 (2021). https://doi.org/10.1038/s41524021005000
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41524021005000
This article is cited by

Complex hexagonal closepacked dendritic growth during alloy solidification by graphics processing unitaccelerated threedimensional phasefield simulations: demo for Mg–Gd alloy
Rare Metals (2023)

Highthroughput discovery of fluorideion conductors via a decoupled, dynamic, and iterative (DDI) framework
npj Computational Materials (2022)

Achieving thermally stable nanoparticles in chemically complex alloys via controllable sluggish lattice diffusion
Nature Communications (2022)

Development of a Diffusion Mobility Database for CoBased Superalloys
Journal of Phase Equilibria and Diffusion (2022)