# Modeling of laser-induced breakdown spectroscopic data analysis by an automatic classifier

- 471 Downloads

## Abstract

Laser-induced breakdown spectroscopy (LIBS) is a multi-elemental and real-time analytical technique with simultaneous detection of all the elements in any type of sample matrix including solid, liquid, gas, and aerosol. LIBS produces vast amount of data which contains information on elemental composition of the material among others. Classification and discrimination of spectra produced during the LIBS process are crucial to analyze the elements for both qualitative and quantitative analysis. This work reports the design and modeling of optimal classifier for LIBS data classification and discrimination using the apparatus of statistical theory of detection. We analyzed the noise sources associated during the LIBS process and created a linear model of an echelle spectrograph system. We validated our model based on assumptions through statistical analysis of “dark signal” and laser-induced breakdown spectra from the database of National Institute of Science and Technology. The results obtained from our model suggested that the quadratic classifier provides optimal performance if the spectroscopy signal and noise can be considered Gaussian.

## Keywords

Spectroscopy Echelle Laser-induced breakdown spectroscopy Optimal classifier Statistical learning theory.## 1 Introduction

Laser-induced breakdown spectroscopy (LIBS) is a multi-elemental and real-time analytical technique with simultaneous detection of all the elements in any type of sample matrix including solid, liquid, gas, and aerosol [1]. In LIBS system, a pulsed laser—such as a Q-switched Nd:YAG, is focused onto the surface of the material to eject a tiny fraction of material (picograms to nanograms) from the surface of the object under investigation. By this process, forming short-lived, highly luminous plasma at the surface of the material is formed. Within this hot plasma, the ejected material is dissociated into excited ionic and atomic species. The excited ions and atoms emit characteristic optical radiation as they revert to lower energy states. Detection and spectral analysis of the optical radiation formed through this process is used to yield information on the elemental composition of the material which includes atomic composition of the compound.

During this excitation process, LIBS not only produces the data associated with the samples of interest but also from the unwanted sources like from the system. LIBS uses multiple spectrograph and synchronized charge-coupled device (CCD) spectral acquisition system to analyze the spectral data. For rapid analysis of heterogeneous materials, the acquisition cycle typically stores 1000 spectra for subsequent filtering and analysis. The incorporation of an effective data analysis methodology has been critical in achieving both accurate and reproducible results in the analysis of samples with the technology. LIBS produces vast amounts of data where one or multiple elements are falling almost at the same emission lines. Simultaneous elemental analysis is required to avoid sampling errors associated with the application of a destructive analysis technique LIBS uses for compositional determination of a heterogeneous material. Simultaneous elemental analysis also reduces the analysis time, thereby increasing sample throughput and efficiency of the whole system. To handle the huge amount of data produced by LIBS, the use of automatic classifier and discriminator for spectral analysis is necessary for accuracy, time saving, and increasing efficiency.

Automatic classification of spectroscopy data is a scientific and technical field where chemical molecules, compounds, and mixtures are distinguished based on their spectral signatures by means of computer algorithms [2, 3]. Automatic classification has been attempted on various spectroscopy techniques: magnetic resonance [4], Fourier transform infrared spectroscopy (FTIR) [5], Raman spectroscopy [6], and LIBS data [7, 8, 9, 10, 11]. The utilized methods usually involve linear models (e.g., linear discriminant analysis [4, 5]) on amplitudes of some spectral components, selected by means of feature selection machine learning algorithms. Other publications describe utilization of principal component analysis of spectral components to reduce data dimensionality, followed by an instance-based machine learning algorithm that provides a linear or nonlinear model [6, 9, 10, 11, 12]. While these approaches may perform well in practice, they are ad hoc and lack theoretical justification; more specifically, there is no assurance of their optimality from the point of view of statistical theory of detection.

## 2 Methodology

### 2.1 Model of spectroscopy system

The intensity of the measured signal is proportional to the echelle efficiency [15, 16] \( e(\lambda )\) that is wavelength dependent.

Note that in the echelle spectrometer, high-order diffraction orders are utilized, and the measurements in each order appear as one linear pattern on the detector. The uneven distribution of orders may lead to closely stacking-up orders and cross talk (“ghost line”) [13, 15]. We model cross talk with a linear system with pulse response \(H_\mathrm{c}\left( \lambda ,\lambda '\right) \).

The light is converted into electrical signal in a CCD sensor, where the number of electrons at each pixel is proportional to the intensity of the incident light at the pixel. In a CCD sensor, three types of noises exist based on the intensity of photon signal present on CCD pixel [17]. These three noises are: read-out noise (at low light intensities), shot noise (at medium intensities), and fixed pattern noise (at high intensities). The shot noise is a combination of photon noise and dark noise. Photon noise comes from random variation of photon flux from the light source, while the dark noise is created because of the thermal generation of carriers. Fixed pattern noise exists because of the variation of charge created in individual pixels of CCD for photon signal input. Considering the laser signal as medium intensity, the dominating noise source for this case is shot noise, which comes mainly from dark current as the device was operating at room temperature. We assume that component of dark current noise is \(n_\mathrm{d}\)(\(\lambda \)) [18]. In the CCD sensor, the signal gets discretized in space (corresponding to discrete wavelength \({\lambda }_k\)) which we model with a low-pass filter \(H_\mathrm{s}\left( \lambda \right) \) followed by multiplication with a Dirac pulse trail \(s_\mathrm{s}\left( \lambda \right) =\sum ^K_{k=1}{\delta \left( \lambda -{\lambda }_k\right) }\) (e.g., [19]). The pixel voltages get amplified (A) and quantized. The amplifier introduces the amplifier noise \(n_\mathrm{a}\)(\({\lambda }_k)\). The quantization adds quantization noise \(n_\mathrm{q}\)(\({\lambda }_k)\). The output of the system is therefore the signal \(s_\mathrm{out}\)(\({\lambda }_k\)) discretized in the wavelength domain. Due to the linearity of the observed system, it can be simplified as shown in Fig. 1b. The output signal \(s_\mathrm{out}\)\({(\lambda }_k\)) consists of the equivalent input signal \(s^{*}_{i}\)\({(\lambda }_k\)) and additive equivalent noise \(n^{*}\)\({(\lambda }_k\)).

Dark current noise \(n_\mathrm{d}\)(\(\lambda )\ \) is here modeled as Gaussian [23]. The read-out noise \(n_\mathrm{a}\)(\({\lambda }_k)\) consists of thermal (Johnson) noise and 1/*f*—noise (flicker) noise and can also be modeled as Gaussian [24]. Quantization noise \(n_\mathrm{q}\)(\({\lambda }_k)\), on the other hand, has uniform distribution (if the quantizer is not overloaded) and is not correlated with the discretized signal value [25]. We assume that the number of quantization levels is large enough so that the influence of \(n_\mathrm{q}({\lambda }_k)\) is small and that \(n^*\left( {\lambda }_k\right) \) can also be modeled as Gaussian.

It is known [26] that dark current noise in CCD detectors is spatially uncorrelated (leading to \(E(n_\mathrm{d}({\lambda }_i)n_\mathrm{d}{(}{\lambda }_i))=0,{\lambda }_i\ne {\lambda }_j\)). (Here, *E* denotes expectation.) We assume that the independence of the noise applies to *all* components of \(n^{*}\)\({(\lambda }_k\)), i.e., \(E({n^{*}}({\lambda }_i)n^*{(}{\lambda }_i))=0,{\lambda }_i\ne {\lambda }_j\).

### 2.2 Optimal classifier of spectroscopy data

*K*-variate Gaussian vectors. By the Gaussian assumption, a sample \({{\mathbf{s}}}_\mathrm{out}\) has the following conditional probability density function under hypothesis \(H_{i}, i=1,2\) [27]:

*K**

*K*covariance matrices are defined as:

- 1.Calculate:where$$\begin{aligned} l\left( {{\mathbf{s}}}_\mathrm{out}\right) ={{\mathbf{s}}}^\mathrm{T}_\mathrm{out}{\mathbf{A}}\varvec{\ }{{\mathbf{s}}}_\mathbf{out }{+}{{\mathbf{b}}}^\mathrm{T}{{\mathbf{s}}}_\mathbf{out } -\gamma \end{aligned}$$(6)$$\begin{aligned} {\mathbf{A}}&\triangleq \frac{1}{2}\left( {\varvec{{\Sigma }}}^{-1}_1-{\varvec{{\Sigma }}}^{-1}_2\right) \nonumber \\ {\mathbf{b}}&\triangleq {\varvec{{\Sigma }}}^{-1}_2{{\mathbf{m}}}_2-{\varvec{{\Sigma }}}^{-1}_1{{\mathbf{m}}}_1 \nonumber \\ \gamma&\triangleq \ln \ \eta +\frac{1}{2}\left( {{\ln } \left| {\varvec{{\Sigma }}}_2\right| \ }-{{\ln } \left| {\varvec{{\Sigma }}}_1\right| }\right. \nonumber \\&\quad \left. {+{{\mathbf{m}}}^\mathrm{T}_2{\varvec{{\Sigma }}}^{-1}_2{{\mathbf{m}}}_2-{{{\mathbf{m}}}^\mathrm{T}_1\varvec{{\Sigma }}}^{-1}_1{{\mathbf{m}}}_1\ }\right) ; \end{aligned}$$(6a)
- 2.
If \(l\left( {{\mathbf{s}}}_\mathrm{out}\right) > 0\), decide \(H_2\); otherwise, decide \(H_1\).

*quadratic*in terms of the observed output vector of the system.

- 1.
Estimate mean vectors \({\mathbf{m}}_1\), \({\mathbf{m}}_2\) and covariance matrices \(\varvec{{\Sigma }}_1^{-1}\), \(\varvec{{\Sigma }}_2^{-1}\) , Eq. (4), from K-dimensional observations data belonging to classes 1 and 2 (and corresponding to \({{H}}_1\), \({{H}}_2\));

- 2.
Choose threshold \(\eta \);

- 3.
Calculate matrix

**A**, vector**b**, and scalar \(\gamma \), Eq. (6a); - 4.
For each sample \(s_\mathrm{out}\), calculate \(l\left( {{\mathbf{s}}}_\mathrm{out}\right) \), Eq. (6), and perform classification.

Observe that the parameters of the decision boundary are not directly related to LIBS wavelengths (but are related to measurements obtained from the spectrometer). In other words, the wavelengths themselves are not input to the model.

**x**and

**y**are

*K*-dimensional feature vectors.

Note that in addition to original features, \(s_\mathrm{out} (\lambda _k), k=1, \dots , K\), the classifier can be applied on *linearly transformed* features \(y_j = f_j(s_\mathrm{out} (\lambda _1 ), \dots , s_\mathrm{out} (\lambda _K ),j = 1, \dots , K'\) where \(K'\le K\). Such features can, e.g., be obtained using principal component analysis (PCA) [29]. In such a case, assuming Eq. (3) holds, the transformed features \(y_j\) also have normal distribution [30]. Hence, for classification of spectroscopy data transformed using PCA, the quadratic classifier is also optimal.

### 2.3 Experimental setup

We utilized Andor Mechelle ME5000 spectrograph with an ICCD camera (iStar, Andor Technology, DH734-18F 03), see Fig. 3. The following parameters are from the technical specifications of the spectrograph and correspond to usual spectroscopy practice. The spectral resolution (the ratio between the wavelength and the smallest difference of wavelengths that can be resolved) was \(R=\) 4000 corresponding to 4 pixels FWHM [31]. The total number of channels was 26,040. The wavelength range was 199.04–974.83 nm. The spectrometer uses diffraction orders \({m} =\) 21–100. The grating had 52.13 line/mm with grating constant \(d\approx \) 5–30 \(\upmu \)m, blazed at 32.35 degree. The spectra were collected 50 ns after the laser pulse with integration time of 700 \(\upmu \)s by an on-board digital delay generator (DDG) of the spectrograph. The CCD was kept at a stable temperature at − 10 \(^{\circ }\)C using a thermoelectric (TE) cooler of the spectrograph to reduce dark signal (see Sect. 2.1). To excite plasma in LIBS [32], a broadband CPA-Series Ti: Sapphire ultra-short laser (Clark-MXR, Inc., Model: 2210) generating 150-fs-long pulses operating at the wavelength of 775 nm was used. For experiments with dark signal, the laser beam was blocked by a nontransparent barrier. This way, we capture only the system’s noise.

## 3 Experimental results

### 3.1 Experiments of “dark signal”

To quantify characteristics of CCD sensor, 1000 dark spectra were acquired with no source of light incident to the sensor. The goal was to test the following hypotheses:

\(\hbox {H}_{01}\): \(s_\mathrm{out}\)(\(\lambda _{k}\)) follows Gaussian distribution, \(\lambda _{k}\in \) [200.33 nm, 909.45 nm].

Note that outside this range the spectrometer provided signals equal to zero for all realizations. The total range of considered wavelengths included 24,650 discrete values.

\(\hbox {H}_{02}\): \(s_\mathrm{out} (\lambda _i)\), \(s_\mathrm{out}(\lambda _j)\) are uncorrelated when \(\lambda _i {\ne }\lambda _j\).

The skewness and kurtosis [35] were calculated for each \(s_\mathrm{out}(\lambda _k)\) using the remaining 992 realizations. Figures 4 and 5 show histograms of the obtained skewness and kurtosis.

*m*, we determined discrete wavelengths \(\lambda _{m,1}<\lambda _{m,2}<\cdots<\lambda _k<\cdots \lambda _{{m,\ m}_k}\) satisfying \(m=\) round \(\left( \frac{20,139}{\lambda _k} \right) \) (where \(\lambda _k\) is given in nanometers) [31]. Then, we computed sample autocorrelations \(r_m (l)\) for signals \(s\left( {\lambda }_{m,1}\right) ,\dots ,s \left( {\lambda }_{{m,m}_k}\right) \) where the signals in each realizations were normalized to have the zero mean. Finally, we averaged normalized correlations \(r_m(l)/r_m(0)\) for \(m=21, \dots 100\). The averaged normalized correlations for lags \(-20, \dots , 20\) are shown in Fig. 6.

### 3.2 Experiments with standardized data

We measured laser-induced breakdown spectroscopy (LIBS) spectra [32] of NIST standardized glass. According to National Institute of Standards & Technology (NIST), the nominal composition of the standard reference wafer 612 used in this work is 72% \(\hbox {SiO}{}_{2}\), 12% CaO, 14% \(\hbox {Na}{}_{2}\)O, and 2% \(\hbox {Al}{}_{2}\hbox {O}{}_{3}\). Total 61 trace elements are included in the glass support matrix. The reference wafer is specifically intended for evaluating analytical techniques used to determine trace elements in inorganic matrices [37]. For a sample of NIST standardized glass, we performed 150 realizations of spectra. This was repeated seven times, for seven different samples, resulting in total of \(n=1050\) spectral realizations. We repeated procedure indicated in Sect. 3.1 to test \(H_{01}\) (Gaussianity). Table 1 indicates spectral ranges where \(H_{01}\) cannot be rejected using different tests and significance levels.

We also computed principal components of the NIST standardized glass data. This resulted in total of \(1049 (=n-1)\) principal components. Out of the first 100 principal components, components 5, 6, 8, 9, 11, 100 were identified to have Gaussian distribution using both Kolmogorov-Smirnov test with \(\alpha = 0.05\) and the Lilliefors test with \(\alpha = 0.005\).

## 4 Discussion

### 4.1 Experiments of “dark signal”

Kolmogorov-Smirnov (KS) test indicates that the hypothesis of Gaussian distribution of “dark signal” holds for almost all wavelengths. (The \(\hbox {H}{}_{01}\) could not be rejected even with very large significance level \(\alpha \).) Lilliefors test indicates that the number of wavelengths where the Gaussian distribution holds is smaller than the number indicated by the KS test. It is shown [38] that KS test tends to be inferior to Lilliefors test when the parameters of the Gaussian distribution are unknown. In such a case, the Lilliefors test has higher power (smaller probability of false acceptance of \(\hbox {H}{}_{01}\)). Hence, there is no wonder that the number of wavelengths where \(\hbox {H}{}_{01 }\) is rejected (Gaussian distribution is not satisfied) is larger with the Lilliefors test than with the Kolmogorov-Smirnov test (for the same \(\alpha =0.05\)).

The estimated normalized autocorrelation of “dark signal,” Fig. 6, indicates that the dark noise samples are observably correlated only with the samples at adjacent wavelengths. Hence, \(\hbox {H}{}_{02}\) cannot be completely accepted. The assumption of whiteness (\(\hbox {H}{}_{02}\)), however, is *not* needed in our model.

### 4.2 Experiments with NIST glass

The results of Kolmogorov-Smirnov and Lilliefors tests on LIBS spectra of standardized NIST glass indicate that the distributions of signals \(s_\mathrm{out}(\lambda _k)\) can be considered approximately Gaussian for a large range of \(\lambda _k\) (notably, when \(\lambda _k \in \) [320 nm, 581 nm] for all attempted tests). Due to observed Gaussian distribution of the dark signal, this leads to conclusion that \(s_i(\lambda _k)\) in the considered case have approximate Gaussian distribution in this range of wavelengths. Furthermore, almost all low-order principal components of the data (that are of practical importance for classification, see e.g., [39]) also have Gaussian distribution.

Spectral ranges where the hypothesis of Gaussianity of LIBS spectra of NIST standardized glass cannot be rejected

Test | Significance level (\(\alpha \)) | Spectral range (nm) |
---|---|---|

Kolmogorov-Smirnov | 0.05 | 306–581 |

Lilliefors | 0.05 | 320–721 |

Lilliefors | 0.005 | 311–749 |

### 4.3 Applicability of the optimal classifier

Classification of LIBS data has been an active area of research. Automatic classification has been attempted on a variety of domains including mineralogy (classification of sedimentary ores [40], quartz samples [41], material science [42], botany [43], homeland security [44], and planetology [45])

The optimal classifier presented in the paper is relatively simple. (Classification is performed by computing a quadratic function of observed discrete spectral components.) This highly contrasts with sophisticated and complex classifiers previously attempted in the literature [7, 9, 10, 11, 42].

The usage of the proposed classifier can be validated using a cross-validation technique [46]. An available dataset is split into *k* disjoint subsets of approximately equal size. The classifier is trained using \(k-1\) subsets, and the classification accuracy is evaluated on the remaining subset. The procedure is repeated until all subsets are utilized for the evaluation of the classifier. This way, assumptions of the optimal classifier can be indirectly validated on particular data. Using this approach, in [39] we demonstrated that a simplified version of the optimal classifier discussed in this study is capable of providing high classification accuracies (\(>90\%\)) when a sufficiently large number of principal components are utilized to perform multi-class classification of LIBS data of four proteins diluted in phosphate-buffered saline solution (bovine serum albumin, osteopontin, leptin, insulin-like growth factor II). This result is in agreement with the findings shown in this study that the principal components predominantly have Gaussian distribution. Note that the applicability of the optimal classifiers depends on our ability to estimate statistical parameters of output signal, Eq. (4), specifically the covariance matrices \(\varvec{{\Sigma }}_i\). If the number of samples per class is small in comparison with the dimension of covariance matrices, additional assumptions about the structure of the matrices are needed (e.g., in [39], we assumed matrix diagonality). Alternatively, the dimensionality of the correlation matrices can be reduced if a number of considered wavelengths are decreased by methods of feature selection (e.g., [35]).

A practitioner may be interested what are the features that are responsible for successful classification. Answer to this depends on which particular classification problem we try to solve (e.g., classification of various compounds, the presence of elements). If feature selection [47] is used for dimensionality reduction, the wavelengths corresponding to selected spectral lines indicate which spectral lines are responsible for building a classification model. In contrast, if feature extraction methods are used [39], the loads (weight factors utilized to calculate principal components) may provide indication of relative feature importance.

By employing support vector machines (SVMs), we can estimate a hypothesis drawn from the function class of polynomials that both separates the data and achieves the maximum margin. SVMs carry the benefit of the descriptive power afforded by models with large degrees of freedom while incurring the complexity (VC dimension) of a relatively small number of support vectors. In the SVM formulation, through “the kernel trick,” a transformation of input space is implemented through the definition of its inner product over the set of in-sample data points. The kernel can be thought of as a transformation of the input space to a high-dimensional representational space (or feature space). This also has the effect of further reducing the computational burden by avoiding computation of inner products in a high-dimensional feature space. The model linear in the feature space induced by the kernel represents the equivalent nonlinear model in the input space. Note that classification of spectroscopy data using SVMs was successfully attempted in [48]. Note, however, that for large *K*, the actual estimation of model coefficients may require excessive computational power.

Equation (6) represents the optimal classifier if the assumption of Gaussian distribution holds. Our experimental results indicate that the Gaussian distribution holds for noise and for *specific* spectroscopy signal in a range of wavelengths. In reality, signals \(s_i\left( \lambda \right) \) have Poisson distribution. If distributions of the signals \(s_i\left( \lambda \right) \) at two different wavelengths are *independent*, the signal components \(\ s_f\left( \lambda \right) \) before sampling will also have Poisson distribution that can be approximated by Gaussian. However, if the distributions are *dependent*, \(s_f\left( \lambda \right) \) as an integral of dependent Poisson variables does not *have* to be Poisson random variable [49]. Further, \(n{}^{*}\)\({(\lambda }_k\)) may not be Gaussian random variables. If the assumptions of Gaussian distribution are not satisfied, techniques of classification of non-Gaussian signals in generalized (non-Gaussian) noise need to be considered [50, 51, 52].

Finally, the optimal classifier presented in this paper assumes that the signal flow in the spectrometer can be represented by *linear* systems \(H_\mathrm{d}\left( \lambda ,\lambda '\right) ,H_\mathrm{c}\left( \lambda ,\lambda '\right) \ \)and \(H_\mathrm{s}\left( \lambda \right) \). Further research is needed to develop the optimal classifier if the assumption of linearity does not hold.

## 5 Conclusions

We discussed the optimal classifier for a signal acquisition model in echelle spectrograph and validated model assumptions in a case of specific LIBS signal. We indicated that the optimal classifier has a quadratic decision boundary and can be approximated using SVMs with a quadratic kernel. The optimal classifier can function with features obtained using a feature selection or feature extraction (principal component analysis) method. Experimental results indicate that in the considered case, the assumptions of Gaussianity hold. Work in progress includes development of the optimal classifier when assumptions of Gaussianity and linearity are relaxed.

## Notes

### Acknowledgements

This work was supported in part by US Department of Defense Breast Cancer Research Program (HBCU Partnership Training Award #BC083639), the US National Science Foundation (CREST Grant #HRD-1242067), CREST-8763, National Aeronautics and Space Administration (URC 7658), and the US Department of Defense/Department of Army (45395-MA-ISP, #54412-CI-ISP, W911NF-11-2-0046). Authors also want to thank Dr. Andrew Maidment and Dr. Predrag Bakic (Univ. Pennsylvania) and Dr. Vojislav Kecman (Virginia Commonwealth Univ).

## References

- 1.Crouch, S., Skoog, D.A.: Principles of Instrumental Analysis. Thomson Brooks/Cole, Australia (2007)Google Scholar
- 2.Götz, M., Kononets, M., Bodenstein, C., Riedel, M., Book, M., Palsson, O.P.: Automatic water mixing event identification in the Koljö Fjord observatory data. Int J Data Sci Anal (2018). https://doi.org/10.1007/s4106
- 3.Weihs, C., Ickstadt, K.: Data science: the impact of statistics. Int. J. Data Sci. Anal.
**6**, 189–194 (2018)CrossRefGoogle Scholar - 4.Nikulin, A.E., Dolenko, B., Bezabeh, T., Somorjai, R.L.: Near-optimal region selection for feature space reduction: novel preprocessing methods for classifying MR spectra. NMR Biomed.
**11**, 209–216 (1998)CrossRefGoogle Scholar - 5.Beleites, C., Steiner, G., Sowa, M.G., Baumgartner, R., Sobottka, S., Schackert, G., Salzer, R.: Classification of human gliomas by infrared imaging spectroscopy and chemometric image processing. Vib. Spectrosc.
**38**, 143–149 (2005)CrossRefGoogle Scholar - 6.Lavine, B.K., Davidson, C.E., Moores, A.J., Griffiths, P.R.: Raman spectroscopy and genetic algorithms for the classification of wood types. Appl. Spectrosc.
**55**, 960–966 (2001)CrossRefGoogle Scholar - 7.Snyder, E.G., Munson, C.A., Gottfried, J.L., De Lucia, F.C., Jr, Gullett B., Miziolek, A.: Laser-induced breakdown spectroscopy for the classification of unknown powders. Appl. Opt.
**47**, G80–G87 (2008)CrossRefGoogle Scholar - 8.Sunku, S., Rao, E.N., Kumar, G.M., Tewari, S.P., Rao, S.V.: Discrimination methodologies using femtosecond LIBS and correlation techniques. Proc. SPIE (2013). https://doi.org/10.1117/12.2015749
- 9.Vance, T., Pokrajac, D., Marcano, A., Markushin, Y., McDaniel, S., Melikechi, N., Lazarevic, A.: Classification of LIBS protein spectra using multi-layer perceptrons. Trans. Mass-Data Anal. Images Signals
**2**, 96–111 (2010)Google Scholar - 10.Pokrajac, D., Vance, T., Lazarevic, A., Marcano, A., Markushin, Y., Melikechi, N., Reljin, N.: Performance of multilayer perceptrons for classification of LIBS protein spectra. In: Proceedings of 10th Symposium Neural Network Applications in Electrical Engineering (NEUREL), Belgrade, Serbia, pp. 171–174 (2010)Google Scholar
- 11.Vance, T., Reljin, N., Lazarevic, A., Pokrajac, D. Kecman, V., Melikechi, N., Marcano, A., Markushin, Y., McDaniel, S.: Classification of LIBS protein spectra using support vector machines and adaptive local hyperplanes. In: Proceedings of 2010 IEEE world congress on computational intelligence, Barcelona, Spain, pp. 1–7 (2010)Google Scholar
- 12.Dharmaraj, S., Jamaludin, A.S., Razak, H.M., Valliappan, R., Ahman, N.A., Harn, G.L., Ismail, Z.: The classification of
*Phyllanthus Niruri*Linn. According to location by infrared spectroscopy. Vib. Spectrosc.**41**, 68–72 (2006)CrossRefGoogle Scholar - 13.Tripathi, M.: Echelle Spectrographs: A Flexible Tool for Spectroscopy: Raman and LIBS Spectroscopy. Andor Technology. http://www.andor.com/pdfs/echelle_spectrograph.pdf (2005). Accessed 06 July 2017
- 14.Palmer, C., Loewen, E.: Diffraction Grating Handbook. Newport Corporation, Rochester (2005)Google Scholar
- 15.Loewen, E., Popov, E.: Diffraction Gratings and Applications. Marcel Dekker Inc., New York (1997)Google Scholar
- 16.Bottema, M.: Echelle efficiency and blaze characteristics. SPIE Proc.
**240**, 171–176 (1981)CrossRefGoogle Scholar - 17.Faraji, K., MacLean, W.J.: CCD noise removal in digital images. IEEE Trans. Image Proc.
**5**, 2676–2685 (2006)CrossRefGoogle Scholar - 18.CCD Image Sensor Noise Sources. Eastman Kodak Company application note MTD/PS-0233, Rochester. https://www.uni-muenster.de/imperia/md/content/ziv/multimedia/downloads/kodak___noise_sources.pdf (2001). Accessed 29 Jan 2019
- 19.Mitra, S.K.: Digital Signal Processing: A Computer-Based Approach. McGraw-Hill, New York (2006)Google Scholar
- 20.Mandel, L.: Fluctuations of photon beams: the distribution of photo-electrons. Proc. Phys. Soc.
**74**, 233–243 (1959)CrossRefGoogle Scholar - 21.Grimmett, G., Welsh, D.: Probability: An Introduction. Oxford Science Publications, Oxford (1986)zbMATHGoogle Scholar
- 22.Haight, F.A.: Handbook of the Poisson Distribution. Wiley, Hoboken (1967)zbMATHGoogle Scholar
- 23.Jain, K.A.: Fundamental of Digital Image Processing. Prentice-Hall, Upper Saddle River (1989)zbMATHGoogle Scholar
- 24.Tian, H.: Noise analysis in CMOS image sensors. Ph.D. dissertation, Stanford University, Stanford, CA (2000)Google Scholar
- 25.Proakis, J.G., Manolakis, D.G.: Digital Signal Processing. Prentice-Hall, Upper Saddle River (1996)Google Scholar
- 26.El Gamal, A., Fowler, B., Min, H., Liu, X.: Modeling and estimation of FPN components in CMOS image sensors. Proc. SPIE Solid State Sens. Arrays Dev. Appl. II
**3301**, 168–177 (1998)Google Scholar - 27.Trees, H.L.V., Bell, K.L.: Detection Estimation and Modulation Theory Part I. Wiley, Upper Saddle River (2013)zbMATHGoogle Scholar
- 28.Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and Other Kernel-based Learning Methods. Cambridge University Press, New York (2000)CrossRefzbMATHGoogle Scholar
- 29.Jolliffe, I.T.: Principal Component Analysis, 2nd edn. Springer, New York (2002)zbMATHGoogle Scholar
- 30.Flury, B.: A: First Course in Multivariate Statistics. Springer, New York (1997)CrossRefzbMATHGoogle Scholar
- 31.Andor\(^{TM}\) Technology: Mechelle User’s Guide. Andor Technology, Belfast (2008)Google Scholar
- 32.Miziolek, A.W., Palleschi, V., Schechater, I.: Laser-Induced Breakdown Spectroscopy (LIBS) Fundamentals and Applications. Cambridge University Press, New York (2006)CrossRefGoogle Scholar
- 33.Massey, F.J.: The Kolmogorov–Smirnov test for goodness of fit. J. Am. Stat. Assoc.
**46**, 68–78 (1951)CrossRefzbMATHGoogle Scholar - 34.Lilliefors, H.W.: On the Kolmogorov–Smirnov test for normality with mean and variance unknown. J. Am. Stat. Assoc.
**62**, 399–402 (1967)CrossRefGoogle Scholar - 35.Dodge, Y.: The Oxford Dictionary of Statistical Terms. Oxford University Press, Oxford (2006)zbMATHGoogle Scholar
- 36.Box, G.E.P., Jenkins, G.M., Reinsel, G.C.: Time Series Analysis: Forecasting and Control. Prentice Hall, Upper Saddle River (1994)zbMATHGoogle Scholar
- 37.https://www-s.nist.gov/srmors/viewTableH.cfm?tableid=90N (2017). Accessed 06 June 2017
- 38.Razali, M., Wah, Y.N.: Power comparisons of Shapiro–Walk, Kolmogorov–Smirnov, Lilliefors and Anderson–Darling tests. J. Stat. Model. Anal.
**2**, 21–33 (2011)Google Scholar - 39.Pokrajac, D., Lazarevic, A., Kecman, V., Marcano, A., Markushin, Y., Vance, T., Reljin, N., McDaniel, S., Melikechi, N.: Automatic classification of laser-induced breakdown spectroscopy (LIBS) data of protein biomarker solutions. Appl. Spectrosc.
**68**, 1067–1075 (2014)CrossRefGoogle Scholar - 40.Pořízka, P., Klus, J., Mašek, J., Rajnoha, M., Prochazka, D., Modlitbová, P., Novotný, J., Burget, R., Novotný, K., Kaiser, J.: Multivariate classification of echellograms: a new perspective in laser-induced breakdown spectroscopy analysis. Sci. Rep.
**7**, 3160 (2017). https://doi.org/10.1038/s41598-017-03426-0 - 41.Ali, A., Khan, M.Z., Rehan, I., Rehan, K., Muhammad, R.: Quantitative classification of quartz by laser induced breakdown spectroscopy in conjunction with discriminant function analysis. J. Spectrosc.
**2016**, 1835027 (2016). https://doi.org/10.1155/2016/1835027 - 42.Zhang, T., Xia, D., Tang, H., Yang, X., Li, Y.: Classification of steel samples by laser-induced breakdown spectroscopy and random forest. Chemom. Intell. Lab. Syst.
**157**, 196–201 (2016)CrossRefGoogle Scholar - 43.Wang, J., Liao, X., Zheng, P., Xue, S., Peng, R.: Classification of Chinese herbal medicine by laser-induced breakdown spectroscopy with principal component analysis and artificial neural network. Anal. Lett.
**51**, 575–586 (2018)CrossRefGoogle Scholar - 44.Hybl, J.D., Lithgow, G.A., Buckley, S.G.: Laser-induced breakdown spectroscopy detection and classification of biological aerosols. Appl. Spectrosc.
**57**, 1207–1215 (2003)CrossRefGoogle Scholar - 45.Lanza, N.L., Wiens, R.C., Clegg, S.M., Ollila, A.M., Humphries, S.D., Newsom, H.E., Barefield, J.E.: Calibrating the ChemCam laser-induced breakdown spectroscopy instrument for carbonate minerals on Mars. Appl. Opt.
**49**, C211–C217 (2010)CrossRefGoogle Scholar - 46.Arlot, S., Celisse, A.: A survey of cross-validation procedures for model selection. Stat. Surv.
**4**, 40–79 (2010)MathSciNetCrossRefzbMATHGoogle Scholar - 47.Liu, H., Motoda, H.: Feature Selection for Knowledge Discovery and Data Mining. Kluwer, Boston (1998)CrossRefzbMATHGoogle Scholar
- 48.Bamgbade, A., Somorjai, R., Dolenko, B., Pranckeviciene, E., Nikulin, A., Baumgartner, R.: Evidence accumulation to identify discriminatory signatures in biomedical spectra. In: Proceedings of 10th conference on artificial intelligence in medicine, 23–27, pp. 463–467 (2005)Google Scholar
- 49.Jacod, J.: Two dependent Poisson processes whose sum is still a Poisson process. J. Appl. Prob.
**12**, 170–172 (1975)MathSciNetCrossRefzbMATHGoogle Scholar - 50.Kletter, D., Schultheiss, P.M., Messer, H.: Optimal detection of non-Gaussian random signals in Gaussian noise. Proc. ICASSP-91
**2**, 1305–1308 (1991)Google Scholar - 51.Nuttall, A.H.: Optimum detection of random signal in non-Gaussian noise for low input signal to noise ratio. Naval Undersea Warfare Center Division. http://www.dtic.mil/dtic/tr/fulltext/u2/a422595.pdf (2017). Accessed 6 July 2017
- 52.Middleton, D.: Non-Gaussian Statistical Communication Theory. Wiley, Hoboken (2012)CrossRefzbMATHGoogle Scholar

## Copyright information

**OpenAccess**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.