Abstract

A scheme for estimating atmospheric parameters Teff, log g and [Fe/H] is proposed on the basis of the Least Absolute Shrinkage and Selection Operator (LASSO) algorithm and Haar wavelet. The proposed scheme consists of three processes. A spectrum is decomposed using the Haar wavelet transform and low-frequency components at the fourth level are considered as candidate features. Then, spectral features from the candidate features are detected using the LASSO algorithm to estimate the atmospheric parameters. Finally, atmospheric parameters are estimated from the extracted spectral features using the support-vector regression (SVR) method. The proposed scheme was evaluated using three sets of stellar spectra from the Sloan Digital Sky Survey (SDSS), Large Sky Area Multi-object Fibre Spectroscopic Telescope (LAMOST) and Kurucz's model, respectively. The mean absolute errors are as follows: for the 40 000 SDSS spectra, 0.0062 dex for log Teff (85.83 K for Teff), 0.2035 dex for log g and 0.1512 dex for [Fe/H]; for the 23 963 LAMOST spectra, 0.0074 dex for log Teff (95.37 K for Teff), 0.1528 dex for log g and 0.1146 dex for [Fe/H]; for the 10 469 synthetic spectra, 0.0010 dex for log Teff (14.42K for Teff), 0.0123 dex for log g and 0.0125 dex for [Fe/H].

INTRODUCTION

The implementation of large-scale and deep sky survey programmes, such as the Sloan Digital Sky Survey (SDSS: York et al. 2000; Ahn et al. 2012), Gaia–ESO Survey (GES: Gilmore et al. 2012; Randich et al. 2013) and Large Sky Area Multi-object Fibre Spectroscopic Telescope (LAMOST/Guoshoujing Telescope: Zhao et al. 2006; Cui et al. 2012), has resulted in the collection of a large number of stellar spectra. The bulk of this stellar spectral information necessitates the utilization of a fully automated characterization process.

In particular, estimating the effective temperature, surface gravity and metallicity from stellar spectra remains a fundamental problem (Wu et al. 2011; Song et al. 2012). For example, Muirhead et al. (2012) investigated the estimation of the effective temperature Teff and metallicity [M/H] for late-K and M-type planet-candidate host stars from the K-band spectra released by the the Kepler mission. In this study, a surface fitting method was used along with three spectral indices, namely the equivalent widths of Na i (2.210 μm) and Ca i (2.260 μm) lines and an index describing the flux change between three 0.02 μm wide bands centred at 2.245, 2.370 and 2.080 μm. Koleva et al. (2009) developed a full-spectrum fitting package, ULySS (University of Lyon Spectroscopic analysis Software), and explored its application in spectral parametrization. Manteiga et al. (2010) parametrized stellar spectra by extracting features based on Fourier analysis and wavelet decomposition as well as by constructing a mapping from feature space to parameter space using a forward neural network (FNN) with three layers.

An automated estimation system for atmospheric parameters from a stellar spectrum is also referred to as a spectrum parametrization system (SPS) in related studies. An SPS system consists of two key processes, namely feature extraction and parameter estimation. Feature extraction determines the representation of spectral information. Principal component analysis (PCA: Re Fiorentin et al. 2007; Bu & Pan 2015) and fast Fourier transforms (FFT: Manteiga et al. 2010) are the two most frequently used feature extraction methods. Parameter estimation constructs a mapping from the extracted features of a spectrum to its atmospheric parameters. Some commonly used estimation methods are the FNN (Re Fiorentin et al. 2007), Gaussian process (GP: Bu & Pan 2015) and support-vector regression (SVR: Li et al. 2014).

This work describes a scheme for estimating atmospheric parameters from the stellar spectrum. The three processes of the proposed scheme are performed as follows. Several candidate features are initially obtained by transforming a spectrum into a low-frequency space using the Haar wavelet transform. Then, some representative candidate features are selected as spectral features using the Least Absolute Shrinkage and Selection Operator (LASSO) algorithm (Tibshirani 1996). Finally, atmospheric parameters are estimated from extracted features using a regression method.

High-frequency components are removed in the first process. These components are usually more affected by noise than low-frequency components, which will be discussed further in Section 5.3. The second process reduces the number of spectral features efficiently, a feature closely related to the estimation efficiency and exploration of representative features (Section 5.3). In the third process, the spectrum parametrization problem is investigated using four typical estimation methods: the radial basis function neural network (RBFNN: Schwenker, Kestler & Palm 2001), SVR (Chang & Lin 2001; Schökopf & Smola 2002; Smola & Schölkopf 2004), K-nearest neighbour regression (KNNR: Altman et al. 1992) and least-squares regression (LSR: James et al. 2013). Experimental results indicate that SVR is superior to the other three methods.

This article is organized as follows. Section 2 describes some applied experimental data. After introducing four regression methods in Section 3, Section 4 describes a scheme for extracting spectral features. In Section 5, the proposed scheme is evaluated experimentally. Finally, Section 6 concludes this work.

DATA

The proposed scheme was evaluated by performing experiments using three data sets: 33 963 actual spectra from LAMOST, 50 000 actual spectra from SDSS and 18 969 synthetic spectra computed from Kurucz's model. Actual spectra usually carry some disturbances from noise and pre-processing imperfections.

The proposed scheme is a statistical learning method. It basically involves the discovery of a mapping from stellar spectra to their atmospheric parameters using empirical data, which is referred to as a training set in machine learning. At the same time, the performance of the mapping discovered should be evaluated objectively. Therefore, an independent set of stellar spectra is needed for this evaluation. This independent set is usually referred to as a test set. Thus, in each experiment, stellar spectra are split into two subsets, namely the training and test sets.

Actual spectra from SDSS

A set of 50 000 stellar spectra and their physical parameters collected from SDSS are selected (Abazajian et al. 2009; Yanny et al. 2009). The selected spectra span the ranges [4088, 9740] K for Teff, [1.015, 4.998] dex for surface gravity and [−3.497, 0.268] dex for [Fe/H]. All stellar spectra are shifted to their rest frames (zero radial velocity) using the radial velocity provided by the SDSS/the Sloan Extension for Galactic Understanding and Exploration (SEGUE) Spectroscopic Parameter Pipeline (SSPP: Beers et al. 2006; Allende Prieto et al. 2008; Lee et al. 2008a,b, 2011; Smolinski et al. 2011) and rebinned to a maximal common log(wavelength) range [3.581862, 3.963961] with a sampling step of 0.0001. The common wavelength range is approximately [3818.23, 9203.67] Å. The SDSS training and test sets are labelled as |$S_{\rm tr}^{\rm SD}$| and |$S_{\rm te}^{\rm SD}$|⁠, respectively. Without special instructions, the sizes of |$S_{\rm tr}^{\rm SD}$| and |$S_{\rm te}^{\rm SD}$| are 10 000 and 40 000, respectively.

Actual spectra from LAMOST

A set of 33 963 LAMOST stellar spectra (Luo et al. 2012) and their physical parameters from the LAMOST pipeline are selected based on two signal-to-noise ratio (SNR) constraints: namely SNRg ≥ 20 and SNRr ≥ 20 in the g and r bands (Luo et al. 2012).1 This constraint is proposed by considering the current quality stability of the LAMOST spectra. The LAMOST spectra are also shifted to their rest frames using the radial velocity provided by LAMOST SSPP and rebinned to their maximal common log(wavelength) range [3.5845, 3.9567] with a sampling step of 0.0001. The common wavelength range is approximately [3841.49, 9051.07] Å. All LAMOST spectra span the ranges [3853.2, 9927] K for Teff, [0.8920, 4.9959] dex for log g and [−2.3280, 0.9360] dex for [Fe/H]. The LAMOST training and test sets are labeled as |$S_{\rm tr}^{\rm LA}$| and |$S_{\rm te}^{\rm LA}$|⁠, respectively. |$S_{\rm tr}^{\rm LA}$| consists of 10 000 stellar spectra, while |$S_{\rm te}^{\rm LA}$| consists of 23 963 spectra.

Synthetic spectra

A set of 18 969 synthetic spectra are calculated from Kurucz's NEWODF (new opacity distribution function) models (Castelli & Kurucz 2003) using the spectrum (v2.76) package (Gray & Corbally 1994) as well as 830 828 atomic and molecular lines (contained in two files, luke.lst and luke.nir.lst). These atomic and molecular data are stored in the file stdatom.dat, which includes solar atomic abundances from Grevesse & Sauval (1998). The spectrum package and the three data files can be downloaded from the website.2

The grids of the synthetic stellar spectra span parameter ranges [4000, 9750] K for Teff (45 values, step sizes of 100 K between 4000 and 7500 K and 250 K between 7750 and 9750 K), [1, 5] dex for log g (17 values, step size of 0.25 dex) and [−3.6, 0.3] dex for [Fe/H] (27 values, step sizes of 0.2 dex between −3.6 and −1 dex and 0.1 dex between −1 and 0.3 dex). The synthetic stellar spectra are also split into two subsets: a training set |$S_{\rm tr}^{\rm SY}$| and a test set |$S_{\rm te}^{\rm SY}$| consisting of 10 469 and 8500 spectra, respectively.

ESTIMATION MODELS AND EVALUATION METHODS

Here, x represents a description of a stellar spectrum, while y is the atmospheric parameter Teff, log g or [Fe/H] of x.

Estimation models

The spectrum parametrization problem is to recover the mapping f:
(1)
from a set of empirical data (training set).

In this work, f is estimated using four typical regression methods: RBFNN (Schwenker et al. 2001), SVR (Chang & Lin 2001; Schökopf & Smola 2002; Smola & Schölkopf 2004), KNNR (Altman et al. 1992) and LSR (James et al. 2013). RBFNN and KNNR are non-linear regression methods, while LSR is a linear regression method. SVR can be implemented with a Gaussian kernel and a linear kernel. These two implementation cases are labelled as SVRG and SVRl, respectively. Therefore, SVRG is a non-linear method whereas SVRl is a linear method.

Evaluation criteria

Suppose that |$\skew4\hat{f}$| is an estimate of the spectrum parametrization mapping f and S = {(x, y)} is a data set, where x is a representation of a stellar spectrum and y is an atmospheric parameter of the corresponding star. The data set S can be a training or test set.

For convenient comparison with related reports, this study evaluates the performance of an estimate |$\skew4\hat{f}$| using three methods: mean error (ME), mean of absolute error (MAE) and standard deviation (SD). These three evaluation methods are widely used in related studies (Re Fiorentin et al. 2007; Jofré et al. 2010; Tan et al. 2013).

Using the data set S, the three evaluation criteria can be defined as follows:
(2)
(3)
(4)
where n is the number of elements in S.

FEATURE EXTRACTION

Extract candidate features

Atmospheric parameters Teff, log g and [Fe/H] show an evident non-linear dependence on stellar spectra (tables 6, 10 and 11 in Li et al. 2014). Therefore, the spectrum parametrization problem is investigated by non-linearly transforming a spectrum before estimating atmospheric parameters. In this work, stellar spectra are transformed using a Haar wavelet (Mallat 2009) and decomposed into a series of components with different wavelengths and frequencies (time–frequency localization).

High-frequency components are usually more affected by noise than low-frequency components. Thus, this work obtains candidate features by removing high-frequency components. This process will be discussed further in Section 5.3.

In addition to the Haar wavelet, multiple alternatives for a non-linear transformation (Mallat 2009; Daubechies 1992), such as the Coiflets, Daubechies, Symmlet and biorthogonal wavelets, are available based on the wavelet transform.

Refining the candidate features

Experiments indicate the presence of many redundancies in the extracted candidate features (Section 5.3). Therefore, this work proposes a scheme for detecting spectral features from the extracted candidate features using the LASSO algorithm (Tibshirani 1996).

Let Str = {(x, y)} be a training set (Section 2), where x = (x1, ⋅⋅⋅, xm) represents a stellar spectrum based on its candidate features, y is an atmospheric parameter of the corresponding star and m is a positive integer. The LASSO algorithm selects features using the following model:
(5)
where λ > 0 is a preset parameter. In this model, only a few |$\hat{w}_i$| will be non-zero and the variables xi with a non-zero |$\hat{w}_i$| are selected as spectral features. The parameter λ controls the number of non-zero parameters |$\hat{w}_i$| or, equivalently, the number of detected features. In this work, the parameter λ is estimated by tenfold cross-validation (Tibshirani 1996; Sjöstrand 2005).

Based on the training set from SDSS (Section 2.1), 17 spectral features are detected for Teff, 24 for log g and 25 for [Fe/H] (Table 1).

Table 1.

Extracted features for estimating atmospheric parameters. WP is the wavelength position represented by a two-dimensional vector [a, b], where a and b are the starting and ending wavelengths (Å).

LabelWP (Å)LabelWP (Å)LabelWP (Å)LabelWP (Å)
(a) Extracted features for estimating Teff.
T1[3932.439, 3946.045]T2[4217.567, 4232.159]T3[4780.375, 4796.915]T4[4851.343, 4868.128]
T5[5033.406, 5050.821]T6[5070.631, 5088.175]T7[5108.131, 5125.804]T8[5126.984, 5144.723]
T9[5145.908, 5163.712]T10[5164.901, 5182.771]T11[5203.098, 5221.100]T12[6562.389, 6585.094]
T13[8524.364, 8553.857]T14[8650.914, 8680.845]T15[8747.058, 8777.321]T16[8844.270, 8874.870]
T17[9008.697, 9039.866]
(b) Extracted features for estimating log g.
L1[3818.229, 3831.440]L2[3889.215, 3902.672]L3[3932.439, 3946.045]L4[4095.076, 4109.245]
L5[4295.978, 4310.841]L6[4540.064, 4555.772]L7[4556.821, 4572.587]L8[4573.640, 4589.464]
L9[4658.671, 4674.789]L10[4833.503, 4850.226]L11[4851.343, 4868.128]L12[4869.249, 4886.096]
L13[4887.221, 4904.130]L14[4923.365, 4940.399]L15[5164.901, 5182.771]L16[5183.964, 5201.900]
L17[5222.302, 5240.371]L18[5241.577, 5259.712]L19[5280.341, 5298.611]L20[5299.831, 5318.167]
L21[5319.392, 5337.796]L22[5418.287, 5437.033]L23[5498.725, 5517.750]L24[6562.389, 6585.094]
(c) Extracted features for estimating [Fe/H].
F1[3932.439, 3946.045]F2[3990.819, 4004.626]F3[4005.549, 4019.407]F4[4020.333, 4034.242]
F5[4035.172, 4049.133]F6[4506.735, 4522.327]F7[4607.465, 4623.406]F8[4745.282, 4761.700]
F9[4780.375, 4796.915]F10[4798.019, 4814.620]F11[4815.729, 4832.390]F12[4851.343, 4868.128]
F13[4869.249, 4886.096]F14[4941.536, 4958.633]F15[4959.775, 4976.935]F16[5051.984, 5069.463]
F17[5108.131, 5125.804]F18[5241.577, 5259.712]F19[5260.924, 5279.126]F20[5280.341, 5298.611]
F21[5299.831, 5318.167]F22[5398.362, 5417.039]F23[5438.285, 5457.101]F24[8524.364, 8553.857]
F25[8650.914, 8680.845]
LabelWP (Å)LabelWP (Å)LabelWP (Å)LabelWP (Å)
(a) Extracted features for estimating Teff.
T1[3932.439, 3946.045]T2[4217.567, 4232.159]T3[4780.375, 4796.915]T4[4851.343, 4868.128]
T5[5033.406, 5050.821]T6[5070.631, 5088.175]T7[5108.131, 5125.804]T8[5126.984, 5144.723]
T9[5145.908, 5163.712]T10[5164.901, 5182.771]T11[5203.098, 5221.100]T12[6562.389, 6585.094]
T13[8524.364, 8553.857]T14[8650.914, 8680.845]T15[8747.058, 8777.321]T16[8844.270, 8874.870]
T17[9008.697, 9039.866]
(b) Extracted features for estimating log g.
L1[3818.229, 3831.440]L2[3889.215, 3902.672]L3[3932.439, 3946.045]L4[4095.076, 4109.245]
L5[4295.978, 4310.841]L6[4540.064, 4555.772]L7[4556.821, 4572.587]L8[4573.640, 4589.464]
L9[4658.671, 4674.789]L10[4833.503, 4850.226]L11[4851.343, 4868.128]L12[4869.249, 4886.096]
L13[4887.221, 4904.130]L14[4923.365, 4940.399]L15[5164.901, 5182.771]L16[5183.964, 5201.900]
L17[5222.302, 5240.371]L18[5241.577, 5259.712]L19[5280.341, 5298.611]L20[5299.831, 5318.167]
L21[5319.392, 5337.796]L22[5418.287, 5437.033]L23[5498.725, 5517.750]L24[6562.389, 6585.094]
(c) Extracted features for estimating [Fe/H].
F1[3932.439, 3946.045]F2[3990.819, 4004.626]F3[4005.549, 4019.407]F4[4020.333, 4034.242]
F5[4035.172, 4049.133]F6[4506.735, 4522.327]F7[4607.465, 4623.406]F8[4745.282, 4761.700]
F9[4780.375, 4796.915]F10[4798.019, 4814.620]F11[4815.729, 4832.390]F12[4851.343, 4868.128]
F13[4869.249, 4886.096]F14[4941.536, 4958.633]F15[4959.775, 4976.935]F16[5051.984, 5069.463]
F17[5108.131, 5125.804]F18[5241.577, 5259.712]F19[5260.924, 5279.126]F20[5280.341, 5298.611]
F21[5299.831, 5318.167]F22[5398.362, 5417.039]F23[5438.285, 5457.101]F24[8524.364, 8553.857]
F25[8650.914, 8680.845]
Table 1.

Extracted features for estimating atmospheric parameters. WP is the wavelength position represented by a two-dimensional vector [a, b], where a and b are the starting and ending wavelengths (Å).

LabelWP (Å)LabelWP (Å)LabelWP (Å)LabelWP (Å)
(a) Extracted features for estimating Teff.
T1[3932.439, 3946.045]T2[4217.567, 4232.159]T3[4780.375, 4796.915]T4[4851.343, 4868.128]
T5[5033.406, 5050.821]T6[5070.631, 5088.175]T7[5108.131, 5125.804]T8[5126.984, 5144.723]
T9[5145.908, 5163.712]T10[5164.901, 5182.771]T11[5203.098, 5221.100]T12[6562.389, 6585.094]
T13[8524.364, 8553.857]T14[8650.914, 8680.845]T15[8747.058, 8777.321]T16[8844.270, 8874.870]
T17[9008.697, 9039.866]
(b) Extracted features for estimating log g.
L1[3818.229, 3831.440]L2[3889.215, 3902.672]L3[3932.439, 3946.045]L4[4095.076, 4109.245]
L5[4295.978, 4310.841]L6[4540.064, 4555.772]L7[4556.821, 4572.587]L8[4573.640, 4589.464]
L9[4658.671, 4674.789]L10[4833.503, 4850.226]L11[4851.343, 4868.128]L12[4869.249, 4886.096]
L13[4887.221, 4904.130]L14[4923.365, 4940.399]L15[5164.901, 5182.771]L16[5183.964, 5201.900]
L17[5222.302, 5240.371]L18[5241.577, 5259.712]L19[5280.341, 5298.611]L20[5299.831, 5318.167]
L21[5319.392, 5337.796]L22[5418.287, 5437.033]L23[5498.725, 5517.750]L24[6562.389, 6585.094]
(c) Extracted features for estimating [Fe/H].
F1[3932.439, 3946.045]F2[3990.819, 4004.626]F3[4005.549, 4019.407]F4[4020.333, 4034.242]
F5[4035.172, 4049.133]F6[4506.735, 4522.327]F7[4607.465, 4623.406]F8[4745.282, 4761.700]
F9[4780.375, 4796.915]F10[4798.019, 4814.620]F11[4815.729, 4832.390]F12[4851.343, 4868.128]
F13[4869.249, 4886.096]F14[4941.536, 4958.633]F15[4959.775, 4976.935]F16[5051.984, 5069.463]
F17[5108.131, 5125.804]F18[5241.577, 5259.712]F19[5260.924, 5279.126]F20[5280.341, 5298.611]
F21[5299.831, 5318.167]F22[5398.362, 5417.039]F23[5438.285, 5457.101]F24[8524.364, 8553.857]
F25[8650.914, 8680.845]
LabelWP (Å)LabelWP (Å)LabelWP (Å)LabelWP (Å)
(a) Extracted features for estimating Teff.
T1[3932.439, 3946.045]T2[4217.567, 4232.159]T3[4780.375, 4796.915]T4[4851.343, 4868.128]
T5[5033.406, 5050.821]T6[5070.631, 5088.175]T7[5108.131, 5125.804]T8[5126.984, 5144.723]
T9[5145.908, 5163.712]T10[5164.901, 5182.771]T11[5203.098, 5221.100]T12[6562.389, 6585.094]
T13[8524.364, 8553.857]T14[8650.914, 8680.845]T15[8747.058, 8777.321]T16[8844.270, 8874.870]
T17[9008.697, 9039.866]
(b) Extracted features for estimating log g.
L1[3818.229, 3831.440]L2[3889.215, 3902.672]L3[3932.439, 3946.045]L4[4095.076, 4109.245]
L5[4295.978, 4310.841]L6[4540.064, 4555.772]L7[4556.821, 4572.587]L8[4573.640, 4589.464]
L9[4658.671, 4674.789]L10[4833.503, 4850.226]L11[4851.343, 4868.128]L12[4869.249, 4886.096]
L13[4887.221, 4904.130]L14[4923.365, 4940.399]L15[5164.901, 5182.771]L16[5183.964, 5201.900]
L17[5222.302, 5240.371]L18[5241.577, 5259.712]L19[5280.341, 5298.611]L20[5299.831, 5318.167]
L21[5319.392, 5337.796]L22[5418.287, 5437.033]L23[5498.725, 5517.750]L24[6562.389, 6585.094]
(c) Extracted features for estimating [Fe/H].
F1[3932.439, 3946.045]F2[3990.819, 4004.626]F3[4005.549, 4019.407]F4[4020.333, 4034.242]
F5[4035.172, 4049.133]F6[4506.735, 4522.327]F7[4607.465, 4623.406]F8[4745.282, 4761.700]
F9[4780.375, 4796.915]F10[4798.019, 4814.620]F11[4815.729, 4832.390]F12[4851.343, 4868.128]
F13[4869.249, 4886.096]F14[4941.536, 4958.633]F15[4959.775, 4976.935]F16[5051.984, 5069.463]
F17[5108.131, 5125.804]F18[5241.577, 5259.712]F19[5260.924, 5279.126]F20[5280.341, 5298.611]
F21[5299.831, 5318.167]F22[5398.362, 5417.039]F23[5438.285, 5457.101]F24[8524.364, 8553.857]
F25[8650.914, 8680.845]

EXPERIMENTS AND DISCUSSION

Performance for SDSS spectra

From the detected features in Table 1, a spectrum parametrization model can be learned from the training set |$S_{\rm tr}^{\rm SD}$| (Section 2.1). The performance obtained using the test set |$S_{\rm te}^{\rm SD}$| is presented in Table 2 .

Table 2.

Performance of the proposed scheme on 40 000 test spectra from SDSS (10 000 SDSS spectra for training, Section 2.1).

Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0065(88.48)4.42 × 10−4(6.28)0.0107(148.04)0.21590.02050.32280.15476.04 × 10−40.2197
SVRG0.0062(85.83)6.05 × 10−4(9.40)0.0101(146.66)0.2035−0.01930.30530.15121.19 × 10−20.2158
KNNR0.0069(94.77)−8.39 × 10−4(−10.13)0.0109(154.62)0.2178−0.03700.30690.2198−3.56 × 10−20.2999
LSR0.0072(99.22)3.45 × 10−4(5.46)0.0111(160.73)0.25940.02700.35740.17863.61 × 10−30.2472
SVRl0.0070(96.77)3.41 × 10−4(7.11)0.0111(162.79)0.24170.04750.36480.1758−8.62 × 10−30.2466
Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0065(88.48)4.42 × 10−4(6.28)0.0107(148.04)0.21590.02050.32280.15476.04 × 10−40.2197
SVRG0.0062(85.83)6.05 × 10−4(9.40)0.0101(146.66)0.2035−0.01930.30530.15121.19 × 10−20.2158
KNNR0.0069(94.77)−8.39 × 10−4(−10.13)0.0109(154.62)0.2178−0.03700.30690.2198−3.56 × 10−20.2999
LSR0.0072(99.22)3.45 × 10−4(5.46)0.0111(160.73)0.25940.02700.35740.17863.61 × 10−30.2472
SVRl0.0070(96.77)3.41 × 10−4(7.11)0.0111(162.79)0.24170.04750.36480.1758−8.62 × 10−30.2466

Notes. The unit for Teff is K; The unit for log Teff is log(K).

Table 2.

Performance of the proposed scheme on 40 000 test spectra from SDSS (10 000 SDSS spectra for training, Section 2.1).

Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0065(88.48)4.42 × 10−4(6.28)0.0107(148.04)0.21590.02050.32280.15476.04 × 10−40.2197
SVRG0.0062(85.83)6.05 × 10−4(9.40)0.0101(146.66)0.2035−0.01930.30530.15121.19 × 10−20.2158
KNNR0.0069(94.77)−8.39 × 10−4(−10.13)0.0109(154.62)0.2178−0.03700.30690.2198−3.56 × 10−20.2999
LSR0.0072(99.22)3.45 × 10−4(5.46)0.0111(160.73)0.25940.02700.35740.17863.61 × 10−30.2472
SVRl0.0070(96.77)3.41 × 10−4(7.11)0.0111(162.79)0.24170.04750.36480.1758−8.62 × 10−30.2466
Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0065(88.48)4.42 × 10−4(6.28)0.0107(148.04)0.21590.02050.32280.15476.04 × 10−40.2197
SVRG0.0062(85.83)6.05 × 10−4(9.40)0.0101(146.66)0.2035−0.01930.30530.15121.19 × 10−20.2158
KNNR0.0069(94.77)−8.39 × 10−4(−10.13)0.0109(154.62)0.2178−0.03700.30690.2198−3.56 × 10−20.2999
LSR0.0072(99.22)3.45 × 10−4(5.46)0.0111(160.73)0.25940.02700.35740.17863.61 × 10−30.2472
SVRl0.0070(96.77)3.41 × 10−4(7.11)0.0111(162.79)0.24170.04750.36480.1758−8.62 × 10−30.2466

Notes. The unit for Teff is K; The unit for log Teff is log(K).

For the SDSS test set, MAE errors are 0.0062, 0.2035 and 0.1512 dex for log Teff, log g and [Fe/H], respectively. To compare the proposed scheme with those in previous related reports, the performance of the proposed scheme was also evaluated using measures ME and SD. More experimental results are presented in Table 2, as well as in Figs 1 and 2. Direct comparisons with reports are shown in Section 6 and more discussion on the dispersion in Fig. 1 is presented in Section 5.4.

Performance of the proposed scheme on 40 000 test spectra from SDSS (10 000 SDSS spectra for training, Section 2.1) using SVRG.
Figure 1.

Performance of the proposed scheme on 40 000 test spectra from SDSS (10 000 SDSS spectra for training, Section 2.1) using SVRG.

Residual distributions of the proposed scheme for 40 000 test spectra from SDSS (10 000 SDSS spectra for training, Section 2.1).
Figure 2.

Residual distributions of the proposed scheme for 40 000 test spectra from SDSS (10 000 SDSS spectra for training, Section 2.1).

Performance for LAMOST spectra and synthetic spectra

The proposed scheme is also tested on actual spectra from LAMOST (Section 2.2) and synthetic spectra (Section 2.3).

The performances of the scheme on LAMOST spectra is presented in Table 3, as well as in Figs 3 and 4. The test results using synthetic spectra are presented in Table 4, as well as in Figs 5 and 6. The results in Tables 23 and 4 show that SVRG and RBFNN are more suitable than KNNR, LSR and SVRl for estimating atmospheric parameters.

Performance of the proposed scheme on 23 963 test spectra from LAMOST (10 000 LAMOST spectra for training, Section 2.2) using SVRG.
Figure 3.

Performance of the proposed scheme on 23 963 test spectra from LAMOST (10 000 LAMOST spectra for training, Section 2.2) using SVRG.

Residual distributions of the proposed scheme for 23 963 test spectra from LAMOST (10000 LAMOST spectra for training, Section 2.2) using SVRG.
Figure 4.

Residual distributions of the proposed scheme for 23 963 test spectra from LAMOST (10000 LAMOST spectra for training, Section 2.2) using SVRG.

Performance of the proposed scheme on 10 469 test spectra computed from Kurucz's model (8500 synthetic spectra for training, Section 2.3) using SVRG.
Figure 5.

Performance of the proposed scheme on 10 469 test spectra computed from Kurucz's model (8500 synthetic spectra for training, Section 2.3) using SVRG.

Residual distributions of the proposed scheme for 10 469 test spectra computed from Kurucz's model (8500 synthetic spectra for training, Section 2.3) using SVRG.
Figure 6.

Residual distributions of the proposed scheme for 10 469 test spectra computed from Kurucz's model (8500 synthetic spectra for training, Section 2.3) using SVRG.

Table 3.

Performance of the proposed scheme on 23 963 test spectra from LAMOST (10 000 LAMOST spectra for training, Section 2.2).

Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0070(91.14)6.75 × 10−5(2.37)0.0099(131.36)0.16640.01090.27530.1197−0.00380.1767
SVRG0.0074(95.37)1.27 × 10−4(4.30)0.0106(141.62)0.1528−0.00080.21020.1146−0.01120.1528
KNNR0.0085(111.80)4.47 × 10−4(10.08)0.0126(173.66)0.19340.01670.27300.1625−0.01540.2151
LSR0.0082(106.51)6.67 × 10−4(11.46)0.0117(161.83)0.22180.01730.34040.1312−0.01430.1807
SVRl0.0081(105.73)1.89 × 10−4(5.44)0.0116(159.83)0.2070−0.01240.31450.1311−0.02220.1806
Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0070(91.14)6.75 × 10−5(2.37)0.0099(131.36)0.16640.01090.27530.1197−0.00380.1767
SVRG0.0074(95.37)1.27 × 10−4(4.30)0.0106(141.62)0.1528−0.00080.21020.1146−0.01120.1528
KNNR0.0085(111.80)4.47 × 10−4(10.08)0.0126(173.66)0.19340.01670.27300.1625−0.01540.2151
LSR0.0082(106.51)6.67 × 10−4(11.46)0.0117(161.83)0.22180.01730.34040.1312−0.01430.1807
SVRl0.0081(105.73)1.89 × 10−4(5.44)0.0116(159.83)0.2070−0.01240.31450.1311−0.02220.1806

Notes. The unit for Teff is K; The unit for log Teff is log(K).

Table 3.

Performance of the proposed scheme on 23 963 test spectra from LAMOST (10 000 LAMOST spectra for training, Section 2.2).

Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0070(91.14)6.75 × 10−5(2.37)0.0099(131.36)0.16640.01090.27530.1197−0.00380.1767
SVRG0.0074(95.37)1.27 × 10−4(4.30)0.0106(141.62)0.1528−0.00080.21020.1146−0.01120.1528
KNNR0.0085(111.80)4.47 × 10−4(10.08)0.0126(173.66)0.19340.01670.27300.1625−0.01540.2151
LSR0.0082(106.51)6.67 × 10−4(11.46)0.0117(161.83)0.22180.01730.34040.1312−0.01430.1807
SVRl0.0081(105.73)1.89 × 10−4(5.44)0.0116(159.83)0.2070−0.01240.31450.1311−0.02220.1806
Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0070(91.14)6.75 × 10−5(2.37)0.0099(131.36)0.16640.01090.27530.1197−0.00380.1767
SVRG0.0074(95.37)1.27 × 10−4(4.30)0.0106(141.62)0.1528−0.00080.21020.1146−0.01120.1528
KNNR0.0085(111.80)4.47 × 10−4(10.08)0.0126(173.66)0.19340.01670.27300.1625−0.01540.2151
LSR0.0082(106.51)6.67 × 10−4(11.46)0.0117(161.83)0.22180.01730.34040.1312−0.01430.1807
SVRl0.0081(105.73)1.89 × 10−4(5.44)0.0116(159.83)0.2070−0.01240.31450.1311−0.02220.1806

Notes. The unit for Teff is K; The unit for log Teff is log(K).

Table 4.

Performance of the proposed scheme on 10 469 test spectra computed from Kurucz's model (8500 synthetic spectra for training, Section 2.3).

Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0010(14.15)1.34 × 10−4(1.42)0.0014(20.26)0.02171.89 × 10−30.05820.02031.95 × 10−30.0282
SVRG0.0010(14.42)2.63 × 10−4(3.34)0.0015(20.81)0.0123−9.42 × 10−40.05900.0125−3.58 × 10−40.0256
KNNR0.0027(39.39)3.92 × 10−5(0.19)0.0041(61.08)0.21673.55 × 10−20.31660.10072.90 × 10−20.1611
LSR0.0026(36.18)−2.54 × 10−4(−3.68)0.0033(46.05)0.14162.36 × 10−20.19020.09039.77 × 10−30.1175
SVRl0.0025(34.91)1.14 × 10−4(1.79)0.0032(45.80)0.13432.21 × 10−20.19200.07834.12 × 10−30.1122
Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0010(14.15)1.34 × 10−4(1.42)0.0014(20.26)0.02171.89 × 10−30.05820.02031.95 × 10−30.0282
SVRG0.0010(14.42)2.63 × 10−4(3.34)0.0015(20.81)0.0123−9.42 × 10−40.05900.0125−3.58 × 10−40.0256
KNNR0.0027(39.39)3.92 × 10−5(0.19)0.0041(61.08)0.21673.55 × 10−20.31660.10072.90 × 10−20.1611
LSR0.0026(36.18)−2.54 × 10−4(−3.68)0.0033(46.05)0.14162.36 × 10−20.19020.09039.77 × 10−30.1175
SVRl0.0025(34.91)1.14 × 10−4(1.79)0.0032(45.80)0.13432.21 × 10−20.19200.07834.12 × 10−30.1122

Notes. The unit for Teff is K; The unit for log Teff is log(K).

Table 4.

Performance of the proposed scheme on 10 469 test spectra computed from Kurucz's model (8500 synthetic spectra for training, Section 2.3).

Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0010(14.15)1.34 × 10−4(1.42)0.0014(20.26)0.02171.89 × 10−30.05820.02031.95 × 10−30.0282
SVRG0.0010(14.42)2.63 × 10−4(3.34)0.0015(20.81)0.0123−9.42 × 10−40.05900.0125−3.58 × 10−40.0256
KNNR0.0027(39.39)3.92 × 10−5(0.19)0.0041(61.08)0.21673.55 × 10−20.31660.10072.90 × 10−20.1611
LSR0.0026(36.18)−2.54 × 10−4(−3.68)0.0033(46.05)0.14162.36 × 10−20.19020.09039.77 × 10−30.1175
SVRl0.0025(34.91)1.14 × 10−4(1.79)0.0032(45.80)0.13432.21 × 10−20.19200.07834.12 × 10−30.1122
Methodlog Teff (Teff)log g[Fe/H]
ParameterMAEMESDMAEMESDMAEMESD
RBFNN0.0010(14.15)1.34 × 10−4(1.42)0.0014(20.26)0.02171.89 × 10−30.05820.02031.95 × 10−30.0282
SVRG0.0010(14.42)2.63 × 10−4(3.34)0.0015(20.81)0.0123−9.42 × 10−40.05900.0125−3.58 × 10−40.0256
KNNR0.0027(39.39)3.92 × 10−5(0.19)0.0041(61.08)0.21673.55 × 10−20.31660.10072.90 × 10−20.1611
LSR0.0026(36.18)−2.54 × 10−4(−3.68)0.0033(46.05)0.14162.36 × 10−20.19020.09039.77 × 10−30.1175
SVRl0.0025(34.91)1.14 × 10−4(1.79)0.0032(45.80)0.13432.21 × 10−20.19200.07834.12 × 10−30.1122

Notes. The unit for Teff is K; The unit for log Teff is log(K).

Filtering and selection: positive or negative?

The proposed scheme extracts spectral features by removing high-frequency components from the Haar wavelet transform and rejecting most low-frequency components by the LASSO algorithm. In this study, we determine whether these processes eliminate important spectral information, such as weak lines.

Four experiments are conducted and the results are listed in Table 5. The results show the possibility of eliminating useful spectral information. However, in application, the observed spectrum is inevitably contaminated with noise. In theory, weak lines should be more sensitive to noise.

Table 5.

In these experiments, the advantages and disadvantages of eliminating high-frequency as well as many low-frequency components are evaluated. The parameters are estimated by a support-vector machine (RBF kernel) and RBF neural network on SDSS samples. Their performances are assessed by MAE. WT(i, 0) and WT(i, 1) represent the coefficients of a wavelet transform with i-level decomposition in the approximation and high-frequency sub-bands. {Ti}, {Li} and {Fi} denote the extracted features for log Teff, log g and [Fe/H], respectively. The number after the colon represents the total number of features utilized.

log Tefflog g[Fe/H]
FeaturesSVRGRBFNNFeaturesSVRRBFNNFeaturesSVRGRBFNN
{Ti}:170.00620.0065{Li}:240.20350.2159{Fi}:250.15120.1547
WT(4,0):2390.00550.0062WT(4,0):2390.19090.2267WT(4,0):2390.13110.1486
WT(4,1)+{Ti}:2560.01650.0083WT(4,1)+{Li}:2630.23680.2449WT(4,1)+{Fi}:2640.18620.1770
Full:38230.04600.0131Full:38230.37260.2366Full:38230.41180.1769
log Tefflog g[Fe/H]
FeaturesSVRGRBFNNFeaturesSVRRBFNNFeaturesSVRGRBFNN
{Ti}:170.00620.0065{Li}:240.20350.2159{Fi}:250.15120.1547
WT(4,0):2390.00550.0062WT(4,0):2390.19090.2267WT(4,0):2390.13110.1486
WT(4,1)+{Ti}:2560.01650.0083WT(4,1)+{Li}:2630.23680.2449WT(4,1)+{Fi}:2640.18620.1770
Full:38230.04600.0131Full:38230.37260.2366Full:38230.41180.1769
Table 5.

In these experiments, the advantages and disadvantages of eliminating high-frequency as well as many low-frequency components are evaluated. The parameters are estimated by a support-vector machine (RBF kernel) and RBF neural network on SDSS samples. Their performances are assessed by MAE. WT(i, 0) and WT(i, 1) represent the coefficients of a wavelet transform with i-level decomposition in the approximation and high-frequency sub-bands. {Ti}, {Li} and {Fi} denote the extracted features for log Teff, log g and [Fe/H], respectively. The number after the colon represents the total number of features utilized.

log Tefflog g[Fe/H]
FeaturesSVRGRBFNNFeaturesSVRRBFNNFeaturesSVRGRBFNN
{Ti}:170.00620.0065{Li}:240.20350.2159{Fi}:250.15120.1547
WT(4,0):2390.00550.0062WT(4,0):2390.19090.2267WT(4,0):2390.13110.1486
WT(4,1)+{Ti}:2560.01650.0083WT(4,1)+{Li}:2630.23680.2449WT(4,1)+{Fi}:2640.18620.1770
Full:38230.04600.0131Full:38230.37260.2366Full:38230.41180.1769
log Tefflog g[Fe/H]
FeaturesSVRGRBFNNFeaturesSVRRBFNNFeaturesSVRGRBFNN
{Ti}:170.00620.0065{Li}:240.20350.2159{Fi}:250.15120.1547
WT(4,0):2390.00550.0062WT(4,0):2390.19090.2267WT(4,0):2390.13110.1486
WT(4,1)+{Ti}:2560.01650.0083WT(4,1)+{Li}:2630.23680.2449WT(4,1)+{Fi}:2640.18620.1770
Full:38230.04600.0131Full:38230.37260.2366Full:38230.41180.1769

Therefore, the loss from elimination is trivial. The wavelet components with the lowest frequency are traditional choices for spectral features for estimating atmospheric parameters (Lu et al. 2013). In the experiments for Teff, when all low-frequency wavelet components are used, the number of features will increase from 17 to 239 (an increase of (239 − 17)/17 = 1305.88 per cent), but the MAE can only decrease by 0.0007 dex (11.29 per cent: Experiments 1 and 2). When no component is eliminated while estimating Teff, the number of features will increase from 17 to 3823 (an increase of 22 388.23 per cent), but the MAE error increases by 0.0398 dex (641.93 per cent). A small number of detected features indicates an efficient process for estimating atmospheric parameters from spectral features. The above results suggest that the model developed in this work can estimate stellar atmospheric parameters with high accuracy.

Knowledge base, dispersion and performance

The proposed scheme is a statistical learning method. Its primary principle is to discover automatically the mapping from a stellar spectrum to its atmospheric parameters from a training set. The training set is the carrier of knowledge and affects the accuracy of the scheme.

Therefore, the size of the training set affects the performance of the proposed scheme. For example, if the size of the training set is increased from 10 000 to 15 000, 20 000 and 25 000 in the experiments on SDSS spectra,3 the test dispersion can clearly be improved (Fig. 7). Similar experiments are conducted on synthetic spectra and the corresponding results are presented in Fig. 8.

Dispersion can be improved by increasing the size of the training set from 10 000 to 15 000, 20 000 and 25 000 in the experiments on SDSS test spectra. This experiment is conducted using SVRG.
Figure 7.

Dispersion can be improved by increasing the size of the training set from 10 000 to 15 000, 20 000 and 25 000 in the experiments on SDSS test spectra. This experiment is conducted using SVRG.

Dispersion can be improved by increasing the size of the training set from 8500 to 10 469 in the experiment on synthetic test spectra. This experiment is conducted using SVRG.
Figure 8.

Dispersion can be improved by increasing the size of the training set from 8500 to 10 469 in the experiment on synthetic test spectra. This experiment is conducted using SVRG.

Actual data usually present some disturbances arising from noise and pre-processing imperfections (e.g. sky lines and/or cosmic-ray removal residuals, residual calibration defects and interstellar extinction instability4). The negative effect from these factors can be reduced to a certain extent by enriching the knowledge carrier, i.e. the training set (Figs 7 and 8).

Compactness

For simplicity, this section considers the features of log g as an example to discuss compactness. Other features of Teff and [Fe/H] can be discussed similarly.

The original SDSS spectra are described by 3821 fluxes. To estimate log g, 24 features are detected and data reduction is (3821 − 24)/(3821) ≈ 99.37 per cent. This result indicates that log g can be estimated from a spectral description with a dimension of 24 instead of 3821. The small number of features also implies high efficiency in estimating the parameter from spectral information.

Compactness indicates the study of whether the number of features can be reduced further. Experimental results show that if seven features, namely L1, L8, L13, L19, L20, L21 and L23, are rejected, the feature number decreases by 7/24 ≈ 29 per cent and the MAE error increases by 0.0027 dex (approximately 0.0027/0.2035 ≈ 1.32 per cent).5 Therefore, the feature number can be refined further if a slight decrease in accuracy is accepted.

CONCLUSION AND FUTURE WORK

This work investigated the estimation of effective temperature (Teff), surface gravity (log g) and metallicity ([Fe/H]) from stellar spectra based on the Haar wavelet transform and LASSO algorithm. The proposed scheme is evaluated using actual spectra from SDSS and LAMOST as well as synthetic spectra computed from Kurucz's model. Favourable results are achieved in all cases.

The proposed scheme exhibits excellent robustness and sparseness. The features are extracted using two steps. From the SDSS data, the original spectra, described by 3821 fluxes, are initially decomposed using the Haar wavelet transform and low-frequency coefficients (239 features) are selected as candidate features. In this step, some noise and redundancies with high-frequency components are removed. Then a small subset of candidate features are chosen as spectral features using the LASSO algorithm. The second step is a supervised learning process that selects features according to their correlation with the parameter to be estimated. The number of selected features is 17 for Teff, 24 for log g and 25 for [Fe/H]. A representative work is Re Fiorentin et al. (2007), in which 50 features are extracted for estimating atmospheric parameters.

Another advantage of the proposed scheme is its high accuracy. Using the SVRG method and 40 000 stellar spectra from SDSS, the MAEs are 0.0062 dex for log Teff (85.83 K for Teff), 0.2035 dex for log g and 0.1512 dex for [Fe/H]. Further details are shown in Table 2. In previous reports, Re Fiorentin et al. (2007) estimated the parameters of 19 000 spectra from SDSS with MAEs of 0.0126 dex for log Teff, 0.3644 dex for log g and 0.1949 dex for [Fe/H]. Jofré et al. (2010) first highly compressed the data using a likelihood method and then estimated the parameters Teff, log g and [Fe/H] from low-resolution stellar spectra measured by SEGUE; the standard deviations (SDs) of the errors were 130 K for Teff, 0.5 dex for log g and 0.25 dex for [Fe/H]. Therefore, the results estimated using the proposed scheme exhibit higher accuracy compared with those reported in the literature.

In this work, the proposed scheme is evaluated using three different data sets (SDSS, LAMOST and synthetic spectra). For each kind of data, the proposed model is learned and tested independently. An interesting problem is the estimation of atmospheric parameters of one data set (e.g. LAMOST) using a model learned from the other data sets (e.g. SDSS or synthetic spectra). For example, Re Fiorentin et al. (2007) investigated how to estimate the atmospheric parameters of SDSS data from synthetic spectra and vice versa. In this process, residual calibration defects should be considered. This work focuses on sparse feature extraction and the abovementioned problem will be investigated in future work.

The authors thank the reviewer and editor for their instructive comments and extend their thanks to Professor Ali Luo and Fang Zuo for their support and discussions. This work is supported by the National Natural Science Foundation of China (grant No: 61273248, 61075033, 61202315), the Natural Science Foundation of Guangdong Province (2014A030313425,S2011010003348), the Open Project Program of the National Laboratory of Pattern Recognition (NLPR) (201001060) and the high-performance computing platform of South China Normal University.

1

This constraint was not used on SDSS data.

3

Correspondingly, the size of the test set decreases from 40 000 to 35 000, 30 000 and 25 000, respectively.

4

By instability, we mean that a slight difference in the interstellar extinction of multiple stars may be observed.

5

These experiments are conducted on SDSS spectra (Section 2.1).

REFERENCES

Abazajian
K. N.
et al.
ApJS
2009
182
543

Ahn
C. P.
et al.
ApJS
2012
203
21

Allende Prieto
C.
et al.
AJ
2008
136
2070

Altman
N. S.
Amer. Stat.
1992
46
175

Beers
T. C.
et al.
Mem. Soc. Astron. Ital.
2006
77
1171

Bu
Y.
Pan
J.
MNRAS
2015
447
256

Castelli
F.
Kurucz
R. L.
Piskunov
N. E.
Weiss
W. W.
Gray
D. F.
IAU Symp. 210, Modelling of Stellar Atmospheres
2003
Cambridge
Cambridge Univ. Press
A20

Chang
C. C.
Lin
C. J.
LIBSVM: A Library for Support Vector Machines
2001

Cui
X.
et al.
Res. Astron. Astrophys.
2012
12
1197

Daubechies
I.
Ten Lectures on Wavelets
1992
Philadelphia
Society for Industrial and Applied Mathematics

Gilmore
G.
et al.
Messenger
2012
147
25

Gray
R. O.
Corbally
C. J.
AJ
1994
107
742

Grevesse
N.
Sauval
A. J.
Space Sci. Rev.
1998
85
161

James
G.
Witten
D.
Hastie
T.
Tibshirani
T.
An Introduction to Statistical Learning with Applications in R
2013
New York
Springer-Verlag

Jofré
P.
Panter
B.
Hansen
C. J.
Weiss
A.
A&A
2010
517
A57

Koleva
M.
Prugniel
P.
Bouchard
A.
Wu
Y.
A&A
2009
501
1269

Lee
Y. S.
et al.
AJ
2008a
136
2022

Lee
Y. S.
et al.
AJ
2008b
136
2050

Lee
Y. S.
et al.
AJ
2011
141
90

Li
X.
Wu
Q. M. J.
Luo
A.
Zhao
Y.
Lu
Y.
Zuo
F.
Yang
T.
Wang
Y
ApJ
2014
790
105

Lu
Y.
Li
X.
Wang
Y.
Yang
T.
Spectrosc. Spectral Anal.
2013
33
2010

Luo
A.
et al.
Res. Astron. Astrophys.
2012
12
1243

Mallat
S.
A Wavelet Tour of Signal Processing
2009
3rd edn
Boston
Academic Press

Manteiga
M.
Ordóñez
D.
Dafonte
C.
Arcay
B.
PASP
2010
122
608

Muirhead
P. S.
Hamren
K.
Schlawin
E.
Rojas-Ayala
B.
Covey
K. R.
Lloyd
J. P.
ApJ
2012
750
L37

Randich
S.
Gilmore
G.
Gaia–ESO Consortium
Messenger
2013
154
47

Re Fiorentin
P.
Bailer-Jones
C. A. L.
Lee
Y. S.
Beers
T. C.
Sivarani
T.
Wilhelm
R.
Allende Prieto
C.
Norris
J. E.
A&A
2007
467
1373

Schölkopf
B.
Smola
A. J.
Learning with Kernels
2002
Cambridge, MA
MIT Press

Schwenker
F.
Kestler
H. A.
Palm
G.
Neural Networks
2001
14
439

Sjöstrand
K.
Matlab Implementation of LASSO, LARS, the Elastic Net and SPCA (Version 2.0), DTU 2005.6
2005
Informatics and Mathematical Modelling
Technical University of Denmark

Smola
A. J.
Schölkopf
B.
Stat. Comput.
2004
14
199

Smolinski
J. P.
et al.
AJ
2011
141
89

Song
Y.
et al.
Res. Astron. Astrophys.
2012
12
453

Tan
X.
Pan
J.
Wang
J.
Luo
A.
Tu
L.
Spectrosc. Spectral Anal.
2013
33
1397

Tibshirani
R.
J. R. Stat. Soc. B
1996
58
267

Wu
Y.
et al.
Res. Astron. Astrophys.
2011
11
924

Yanny
B.
et al.
AJ
2009
137
4377

York
D. G.
et al.
AJ
2000
120
1579

Zhao
G.
Chen
Y.
Shi
J.
Liang
Y.
Hou
J.
Chen
L.
Zhang
H.
Li
A.
Chin. J. Astron. Astrophys.
2006
6
265