Abstract
In this paper, deep back propagation neural networks (DBP-NNs) and radial basis function neural networks (RBF-NNs) are employed to predict the dispersion relations (DRs) of one-dimensional (1D) phononic crystals (PCs). The data sets generated by transfer matrix method (TMM) are used to train the NNs and detect their prediction accuracy. In our work, filling fractions, mass density ratios and shear modulus ratios of PCs are considered as the input values of NNs. The results show that both the DBP-NNs and the RBF-NNs exhibit good performances in predicting the DRs of PCs. For one-parameter prediction, the RBF-NNs have shorter training time and remarkable prediction accuracy, for two- and three-parameter prediction, the DBP-NNs have more stable performance. The present work confirms the feasibility of predicting the DRs of PCs by NNs, and provides a useful reference for the application of NNs in the design of PCs and metamaterials.
Similar content being viewed by others
Introduction
A composite structure in which elastic constants are periodically distributed is called a phononic crystal (PC). When an elastic wave is periodically modulated by the elastic constant, a phonon bandgap may be generated, and the propagation of elastic waves in a certain frequency range is suppressed or prohibited. This bandgap characteristics opened a new prospect in wave control and manipulations1.
In dealing with the calculation of PCs, especially with the structure optimization, the traditional calculation methods, such as transfer matrix method (TMM), plane wave expansion method, finite element method, etc., consume a lot of time and computer memory.
Benefiting from the concept of deep learning method2, artificial intelligence technology has ushered in a rapid development in the past few years. As an important member of artificial intelligence technology, neural network (NN) plays an important role in all walks of life, owing to its massively parallel distributed structure of NN and its superior learning ability. Google’s “Alpha Go”, autopilot technology, face recognition technology, text translation technology and medical automatic diagnosis technology, etc., are typical examples of NN applications.
PCs were proposed by analogy to photonic crystals3,4,5,6, but are more complex because of the vector property of elastic waves. Few literatures have been mentioned so far involving the application of NNs in PCs except for the work by Finol et al.7 who compared the prediction accuracy of multilayer perceptron and convolutional neural network (CNN) for one-dimensional (1D) PCs; meanwhile, he used CNN to predict the eigenvalues of two-dimensional (2D) PCs. However, in the field of photonic crystal, there have been some preliminary results. Adriano et al.8,9 used a multilayer perceptron and an extreme learning machine to identify the bandgap width of 2D photonic crystals and the dispersion curves of 2D and three-dimensional (3D) photonic crystals. Christian et al.10 used a NN with a hidden layer to identify the crystalized size of the zinc oxide quantum dots and the energy of the band gap. Liu et al.11 achieved the inverse design of optical wave metamaterials by generative adversarial networks. Dong et al.12 compared the performance of three convolutional NNs and a support vector machine to identify the photonic bandgap width of a 2D “graphene-boron nitride” periodic structure.
This study aims to predict the DRs of 1D PCs and reduce the calculation cost by means of NNs. Two NN models, deep back propagation neural network (DBP-NN) and radial basis function neural network (RBF-NN)13,14, are trained, and their performances are tested. The investigation has been carried out for three cases, that are respectively one-, two- and three-parameter prediction. The present work is hopeful to provide an effective method for the analysis of PCs and to lay a foundation for the intelligent inverse design of PCs and metamaterials.
Problem Description
Consider shear-horizontal (SH) waves propagate in a 1D PC shown in Fig. 1. Material A and B are periodically arranged in x-direction with \(\,a={a}_{A}+{a}_{B}\), where a is a lattice constant; aA and aB are, respectively, the thickness of material A and B in one-unit cell; and θ denotes the incident direction of SH waves.
It is known that the TMM has advantages in dealing with 1D PCs. When the SH wave is obliquely incident on the 1D PCs, its governing equation can be expressed as
where \(u(x,z,t)\) is displacement of the PCs in the y direction, μ is shear modulus, \(\rho \) is mass density, and t is time.
When a harmonic plane wave is considered, it can be assumed
where \(\omega \) is angular frequency, and kz is wave number of the SH wave in the z direction, which is a constant from the Snell theorem. Substituting Eq. (2) into Eq. (1), the displacement and the stresses of the material A and the material B in the n-th lattice are respectively
where, for \(i=A\), \((n-1)a\le x\le (n-1)a+{a}_{A}\), for \(i=B\), \((n-1)a+{a}_{A}\le x\le na\), and
ci and αi (\(i=A,B\)) are the wave velocity and the wave vector components in x direction respectively, and \({C}_{ni}^{tr}\) and \({C}_{ni}^{r}\) (\(i=A,B\)) indicate the amplitudes of the transmitted and reflected waves of the material i in the n-th lattice respectively.
The displacement and stress at the interface between material A and material B are continuous, so when \(x=(n-1)a\), \({u}^{nA}={u}^{(n-1)B}\) and \({\tau }_{xy}^{nA}={\tau }_{xy}^{(n-1)B}\), and when \(x=(n-1)a+{a}_{A}\), \({u}^{nA}={u}^{nB}\) and \({\tau }_{xy}^{nA}={\tau }_{xy}^{nB}\).
Set
The relationship between the amplitudes of the incident and the reflected waves in the n-th lattice and the (n − 1)-th lattice can be obtained from Eq. (4), that is
where T is the transfer matrix, and \({\boldsymbol{T}}={{\boldsymbol{K}}}_{2}^{-1}{{\boldsymbol{H}}}_{2}{{\boldsymbol{H}}}_{1}^{-1}{{\boldsymbol{K}}}_{1}\).
Due to the periodicity in the x direction, use the Bloch theorem to get
where k is a 1D Bloch wave vector.
Substituting Eq. (5) into Eq. (6) gives a standard matrix eigenvalue problem, that is
where I is a 2 × 2 unit matrix.
By solving the eigenvalues of the matrix T, the DRs between the wave vector k and the angular frequency \(\omega \) can be obtained14
If the material A and B are chosen, respectively, as aluminum and epoxy; \({a}_{A}={a}_{B}=0.075\,{\rm{m}}\), and \(\theta =0.5\,{\rm{rad}}\); the DRs of a PC, i.e., the relations between eigen frequency and wave vector, can be obtained according to Eq. (8). Figure 2 gives the first three eigen modes of the PC, where \(\Omega =\omega a/2\pi {c}_{A}\), is the normalized frequency.
From Eq. (8), we can see, the DRs of 1D PCs are determined by \({a}_{i}\), \({\mu }_{i}\) and \({\rho }_{i}(i=A,B)\) for a certain incident wave and periodic constant. Considering that for anti-plane waves, the shear modulus ratio, \(\bar{\mu }={\mu }_{B}/{\mu }_{A}\), and the mass density ratio, \(\bar{\rho }={\rho }_{B}/{\rho }_{A}\) are the main physical parameters affecting the PC bandgap15, and the filling fraction, \(\bar{a}={a}_{B}/a\), is the key geometrical parameter. Hence, for our problem, we focus on the study of parameters of \(\bar{\mu }\), \(\bar{\rho }\) and \(\bar{a}\) respectively. Thus, the relationship between DRs and PCs parameters can be written as
where drs represents the DRs, and fa is the function which reflects the analytic relations between \(\bar{a}\), \(\bar{\mu }\), \(\bar{\rho }\) and drs. The function fa belongs to the category of transcendental equations, so it is difficult to get analytical solutions.
In this paper, our aim is training NNs to learn the relationship between DRs and PCs parameters or, in other words, making NNs simulate the function fa. The simulating relationship is written as
where fs is the function which simulates the function fa. The ultimate goal of our work is obtaining the function fs by NN.
We investigate this simulating relationship between DRs and PCs parameters for three cases. First, we investigate the relationship between \(\bar{a}\) and DRs, then the relationship between \(\bar{\mu }\), \(\bar{\rho }\) and DRs, and finally, the relationship between \(\bar{a}\), \(\bar{\mu }\), \(\bar{\rho }\) and DRs.
Neural Networks and Data Set
In our work, the key issue is to get an input-output relationship by NN, where the input is the PCs parameters, namely, \(\bar{a}\), \(\bar{\mu }\) and \(\bar{\rho }\), and the output is the DRs. NN is a technology driven by data. It can learn the features of input-output through being fed enough data. The NN returned from learning can predict the corresponding output, if an input is given, which is never learned by the NN. We are going to use enough data to train NNs for making them simulate the relationship between DRs and PCs parameters.
DBP-NN
The back propagation (BP) NN generally refers to a feedforward NN. The number of the hidden layers in the BP-NN was usually only three layers at most at the past. However, in our work, the number of the hidden layers is more than three layers, so the BP-NNs we use are deeper in the hidden layers. Hence, we call them DBP-NNs. The deeper the hidden layers, the stronger the ability to learn the features of data. The structure of a DBP-NN is shown in Fig. 3, which consists of an input layer, multiple hidden layers and an output layer. Each layer is composed of many neurons except for input layer. Each neuron consists of inputs, weights, a bias, and an activation function. The output of each neuron is defined as
where x is the inputs, w is the weights, b is the bias, fo is the activation function, and yo is the output of the neuron. A satisfied NN can be obtained by adjusting the weights and biases.
Cost function
It has been proved that NN, whose cost function is mean squared error (MSE) function, can estimate posterior probabilities for finite samples with high accuracy16,17. For the DBP-NNs, we take the MSE function as the cost function, which is defined as
where n is the number of groups of training set data, y is the target outputs (DRs in this paper), and yNN is the outputs of DBP-NN during training.
RBF-NN
Radial basis function neural network (RBF-NN) can approximate arbitrary nonlinear functions. With good generalization ability, it is able to learn complicated laws in a system, and its learning efficiency is remarkable. RBF-NN is composed of an input layer, a hidden layer and an output layer, as is shown in Fig. 4. For the RBF-NN, there is no weight connecting between the input layer and the hidden layer, but the weight connecting between the output layer and the hidden layer. RBFs can calculate the distance or similarity between the inputs and the centers of the hidden layer. The farther the distance is or the lower the similarity is, the smaller the activation of a neuron is, and the less obvious its effect is.
Linear regression
For the RBF-NNs in this paper, we use linear regression method to calculate the weight between the hidden layer and the output layer. Compared to the gradient descent method, linear regression method saves the training time and its model is simpler. The weights calculated by linear regression method are as follows
where β is the weights, y is the target outputs, and yNN is the outputs of RBF-NN during training.
Data set
The data set is composed of the training set, the validation set and the testing set, where the data in the training set, the validation set and the testing set are completely different from each other. The parameters, filling fraction \(\bar{a}\), shear modulus ratio \(\bar{\mu }\) and mass density ratio \(\bar{\rho }\), are taken as the input of the NNs, and the first three eigen modes of the corresponding DRs calculated by TMM as the labels. In our work, three cases are considered. For the first case, the training set A, validation set A and testing set A consist of 10, 2 and 2 sets of data respectively, where \(\bar{\mu }\) and \(\bar{\rho }\) are unchanged and the range of \(\bar{a}\) is from 0.3 to 0.75; for the second case, the training set B, validation set B and testing set B consist of 100, 20 and 20 sets of data respectively, where \(\bar{a}\) is unchanged and the ranges of \(\bar{\mu }\) and \(\bar{\rho }\) are respectively from 0.005 to 0.095 and from 0.1667 to 0.5667; for the third case, the training set C, validation set C and testing set C consist of 1000, 100 and 100 sets of data respectively, where the range of \(\bar{a}\), \(\bar{\mu }\) and \(\bar{\rho }\) are respectively from 0.3 to 0.75, from 0.005 to 0,095 and from 0.1667 to 0.5667.
Results and Discussions
The performances of the trained DBP-NNs and RBF-NNs are tested for three cases, involving geometric parameter changes, physical parameters changes and simultaneous changes, respectively.
The computing platform used in our work is a laptop whose configuration is shown in Table 1.
All the programs are written in “Python 3.5”. The DBP-NNs are developed in “Tensorflow”. The function of “time.clock()” is used to calculate the running time of programs.
Here we measure the prediction accuracy by calculating the Euclidean distance (ED) between the predicted DRs and the target DRs. The smaller the Euclidean distance is, the higher the prediction accuracy is. The ED is defined as
where, yi is the target value, \({\tilde{y}}_{i}\) is the predicted value, and n is the dimension of DRs.
The choice of NNs architectures
In this section, the architectures of the two NNs are discussed for three cases. The optimal architectures of the two NNs are determined by comparing the mean errors of the corresponding validation sets.
Choosing the DBP-NNs architectures
For one-parameter prediction, the following four architectures of the DBP-NNs are compared:
DBP-1-1: 1-30-30-30-303
DBP-1-2: 1-30-30-30-30-303
DBP-1-3: 1-30-30-30-30-30-303
DBP-1-4: 1-30-30-30-30-30-30-303
For two-parameter prediction, the following four architectures of the DBP-NNs are compared:
DBP-2-1: 2-300-100-100-100-303
DBP-2-2: 2-300-100-100-100-100-303
DBP-2-3: 2-300-100-100-100-100-100-303
DBP-2-4: 2-300-100-100-100-100-100-100-303
For three-parameter prediction, the following four architectures of the DBP-NNs are compared:
DBP-3-1: 3-600-300-100-100-303
DBP-3-2: 3-600-300-100-100-100-303
DBP-3-3: 3-600-300-100-100-100-100-303
DBP-3-4: 3-600-300-100-100-100-100-100-303
where “1”, “2” or “3” is the dimension of the input layer, “303” is the number of the neurons in the output layer, and others are the number of the neurons in the hidden layers.
Figure 5 gives the mean errors of the validation sets of the DBP-NNs under three cases. It can be seen that for one-parameter prediction, the mean error of “DBP-1-3” is smaller than others, for two-parameter design, “DBP-2-3” and “DBP-2-4” have a similar accuracy, but “DBP-2-3” has less hidden layers than “DBP-2-4”, and for three-parameter design, “DBP-3-3” is the best choice.
Choosing the RBF-NNs architectures
Similar to the previous section, for one-, two- and three-parameter prediction, the following twelve architectures of the RBF-NNs are respectively compared:
RBF-1-1: 1-5-303
RBF-1-2: 1-10-303
RBF-1-3: 1-15-303
RBF-1-4: 1-20-303
RBF-2-1: 2-50-303
RBF-2-2: 2-100-303
RBF-2-3: 2-150-303
RBF-2-4: 2-200-303
RBF-3-1: 3-200-303
RBF-3-2: 3-400-303
RBF-3-3: 3-600-303
RBF-3-4: 3-800-303
Figure 6 gives the mean errors of the validation sets of the RBF-NNs under three cases. It can be seen that “RBF-1-3”, “RBF-2-3” and “RBF-3-3” are the optical architectures for the first, second and third cases, respectively.
One-parameter prediction
In this section, the DRs of PCs with different filling fractions are predicted, and the data set is composed of the training set A, the validation set A and the testing set A. In this section, the architecture of the DBP-NN is “DBP-1-3”, and the architecture of the RBF-NN is “RBF-1-3”. The predicted DRs of the testing set A are shown in Fig. 7, and the prediction accuracies are shown in Fig. 8. It can be seen that the predicted DRs are in good agreement with the target values. The two NNs exhibit good performances for predicting the DRs of PCs with different filling fractions, but the RBF-NN is better.
Two-parameter prediction
In this section, the DRs of PCs with different shear modulus ratios and mass density ratios are predicted, and the data set is composed of the training set B, the validation B and the testing set B. Due to space limitations, only 2 sets of predictions are shown in Fig. 9 as examples, but the prediction accuracies of the 20 sets are all shown in Fig. 10. Here, the architecture of the DBP-NN is “DBP-2-3”, and the architecture of the RBF-NN is “RBF-2-3”. It can be noticed from Fig. 9 that for two-parameter, both the two NNs present satisfied predictions with high precision. From Fig. 10, it can be seen the DBP-NN has more stable performance, and the RBF-NN has several relatively large errors, although its most errors are very small.
Three-parameter prediction
Predictions of the DRs for different filling fractions, shear modulus ratios and mass density ratios are carried out, and the data set is composed of the training set C, the validation set C and the testing set C. Only 2 sets of predictions are shown in Fig. 11 as examples, while the error statistics of the predicted results of the testing set C are shown in Fig. 12. In this section, the architecture of the DBP-NN is “DBP-3-3”, and the architecture of the RBF-NN is “RBF-3-3”. It can be seen that the performances of the two NNs are still remarkable, but the DBF-NN performs much better.
Comparison among DBP-NNs, RBF-NNs and TMM are given in Table 2. It can be noticed that the time required by NNs is extremely short, and the prediction accuracy is remarkable. For one-parameter prediction, the RBF-NNs are superior to the DBP-NNs on training time, prediction accuracy and simplicity of the model. For two-parameter prediction, the RBF-NN has a smaller mean error, but the DBP-NNs are better than the RBF-NNs in terms of the performance stability from Fig. 10. For three-parameter prediction, the DBP-NN is a better choice because of its high accuracy and stability. In terms of calculation time, TMM is hundreds of thousands of times DBP-NNs and RBF-NNs
Conclusions
The deep back propagation neural networks (DBP-NNs) and the radial basic function neural networks (RBF-NNs) are trained to predict the dispersion relations (DRs) of one-dimensional (1D) phononic crystals (PCs) for three different cases in our work. The results show that both the DBP-NNs and the RBF-NNs can predict the DRs of PCs with rather short time and high accuracy. For one-parameter prediction, the RBF-NNs are superior to the DBP-NNs on training time, prediction accuracy and simplicity of the model. For two-parameter prediction, the DBP-NN has more stable performance. For three-parameter prediction, the DBP-NN is a better choice because of its high accuracy and stability.
This paper confirms the feasibility and superiority of using NNs to predict the DRs of PCs. It is the fact that 2D and 3D problems are more complex in design and calculation, consuming much more time and computer memory. Therefore, the application of NNs to the design and analysis of 2D and 3D PCs and metamaterials will be of great significance. How to design a suitable NN to solve the problems of PCs and metamaterials, especially their inverse design problems, will be the difficulty and focus of the future research. The present work provides a useful reference for the related investigations in the future.
Change history
15 January 2020
An amendment to this paper has been published and can be accessed via a link at the top of the paper.
References
Chaunsali, R., Li, F. & Yang, J. Stress Wave Isolation by Purely Mechanical Topological Phononic Crystals. Scientific Report 6, 30662 (2016).
LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436 (2015).
Yablonovitch, E. Inhibited spontaneous emission in solid-state physics and electronics. Phys. Rev. Lett. 58, 2059–2062 (1987).
John, S. Strong localization of photons in certain disordered dielectric superlattices. Phys. Rev. Lett. 58, 2486–2489 (1987).
Sigalas, M. M. & Economou, E. N. Elastic and acoustic wave band structure. J. Sound Vib. 158, 377–382 (1992).
Kushwaha, M. S. et al. Acoustic band structure of periodic elastic composites. Phys. Rev. Lett. 71, 2022–2025 (1993).
Finol, D., Lu, Y., Mahadevan, V. & Srivastava, A. Deep convolutional neural networks for eigenvalue problems in mechanics. International Journal for Numerical Methods in Engineering, https://doi.org/10.1002/nme.6012 (2018).
Ferreira, A. D. S., Silveira, G. N. M. & Figueroa, H. E. H. Predicting complete band-gaps of 2d photonic crystals by using artificial neural networks. International Microwave and Optoelectronics Conference (IMOC) on 1–5, https://doi.org/10.1109/IMOC.2017.8121135 (IEEE, 2017).
Ferreira, A. D. S., Silveira, G. N. M. & Figueroa, H. E. H. Computing optical properties of photonic crystals by using multilayer perceptron and extreme learning machine. Journal of Lightwave Technology 36, 4066–4073 (2018).
Pelicano, C. M., Rapadas, N., Cagatan, G. & Magdaluyo, E. Jr. Application of back-propagation artificial neural network (ANN) to predict crystallite size and band gap energy of zno quantum dots. AIP Conference Proceedings 1901, 020014 (2017).
Liu, Z., Zhu, D., Rodrigues, S. P., Lee, K. T. & Cai, W. A generative model for inverse design of metamaterials. Nano Letters 18, 6570–6576 (2018).
Dong, Y. et al. Deep learning bandgaps of topologically doped graphene. arXiv preprint arXiv:1809.10860 (2018).
Moody, J. & Darken, C. J. Learning with localized receptive fields. (Morgan Kaufmann Publishers, 1988).
Camley, R. E., Rouhani, B. D., Dobrzynski, L. & Maradudin, A. A. Transverse elastic waves in periodically layered infinite and semi-infinite media. Phys. Rev. B 27, 7318–7329 (1983).
Zhou, X. Z., Wang, Y. S. & Zhang, C. Effects of material parameters on elastic band gaps of two-dimensional solid phononic crystals. Journal of Applied Physics 106, 014903 (2009).
Richard, M. D. & Lippmann, R. P. Neural network classifiers estimate Bayesian a posteriori probabilities. Neural Comput. 3, 461–483 (1991).
Hung, M. S., Hu, M. Y., Shanker, M. S. & Patuwo, B. E. Estimating posterior probabilities in classification problems with neural networks. Int J Comput Intell Organ. 1, 49–60 (1996).
Acknowledgements
This work is supported by the National Natural Science Foundation of China (No. 11772040).
Author information
Authors and Affiliations
Contributions
Chen-Xu Liu drafted the manuscript, performed all calculations and conducted physical analysis. Gui-Lan Yu supervised the project and contributed to the writing and revision. All authors discussed the results and contributed to the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Liu, CX., Yu, GL. Predicting the Dispersion Relations of One-Dimensional Phononic Crystals by Neural Networks. Sci Rep 9, 15322 (2019). https://doi.org/10.1038/s41598-019-51662-3
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-019-51662-3
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.