Abstract

The advancement of technology has resulted in a rapid increase in supernova (SN) discoveries. The Subaru/Hyper Suprime-Cam (HSC) transient survey, conducted from fall 2016 through spring 2017, yielded 1824 SN candidates. This gave rise to the need for fast type classification for spectroscopic follow-up and prompted us to develop a machine learning algorithm using a deep neural network with highway layers. This algorithm is trained by actual observed cadence and filter combinations such that we can directly input the observed data array without any interpretation. We tested our model with a dataset from the LSST classification challenge (Deep Drilling Field). Our classifier scores an area under the curve (AUC) of 0.996 for binary classification (SN Ia or non-SN Ia) and 95.3% accuracy for three-class classification (SN Ia, SN Ibc, or SN II). Application of our binary classification to HSC transient data yields an AUC score of 0.925. With two weeks of HSC data since the first detection, this classifier achieves 78.1% accuracy for binary classification, and the accuracy increases to 84.2% with the full dataset. This paper discusses the potential use of machine learning for SN type classification purposes.

1 Introduction

Time domain science has become a major field of astronomical study. The discovery of the accelerating universe (Riess et al. 1998; Perlmutter et al. 1999) invoked a series of large supernova (SN) surveys in recent decades (Betoule et al. 2014; Scolnic et al. 2018; Brout et al. 2019). These surveys revealed that an entirely new family of transients exists, and the search for unknown populations is currently of great interest (Howell et al. 2006; Phillips et al. 2007; Quimby et al. 2007). For precision cosmology, it is highly important to maintain the purity of Type Ia supernovae (SNe Ia) while performing uniform sampling of these SNe ranging from bright to faint. Spectroscopic follow-up has been essential to distinguish a faint SN Ia from a Type Ib supernova (SN Ib) and a Type Ic supernova (SN Ic), which have similar light curve behavior (Scolnic et al. 2014). Considering that a large amount of precious telescope time is dedicated to these follow-up programs, it is desirable to make efficient use of these telescopes.

The scale of surveys is becoming larger, and it has become impossible to trigger spectroscopic follow-up for all of the candidates in real time. It has therefore become necessary to develop a new classification scheme, and a natural path along which to proceed would be to perform photometric classification (Sako et al. 2011; Jones et al. 2018). The rise of machine learning technologies has resulted in astronomical big data commonly being analyzed using machine learning techniques.

Neural networks have been used for photometric redshift studies from the early stages. Today, many machine learning methods are applied to photometric redshift studies (Collister & Lahav 2004; Carliles et al. 2010; Pasquet et al. 2019). Deep neural networks (DNNs) are being used to process imaging data to identify strong lens systems (Petrillo et al. 2017) or for galaxy morphology classifications (Hausen & Robertson 2020). Currently, machine learning is being introduced to process transient survey data for detection (Goldstein et al. 2015) and classification purposes (Charnock & Moss 2017). A recurrent autoencoder neural network (RAENN) is introduced for photometric classification of the SN light curves (Villar et al. 2020), and being applied to 2315 Pan-Starrs1 data (Hosseinzadeh et al. 2020). In preparation for the Vera C. Rubin Observatory, a real timeclassifier, ALeRCE (Automatic Learning for the RapidClassification of Events), is being developed (Sánchez-Sáezet al. 2020; Förster et al. 2020) and currently applied to the Zwicky Transient Facility (ZTF) data (Carrasco-Davis et al. 2020).

In this paper we introduce our attempt to apply machine learning to actual transient survey data. The Hyper Suprime-Cam (HSC; Miyazaki et al. 2018; Komiyama et al. 2018; Kawanomoto et al. 2018; Furusawa et al.

2018), a gigantic mosaic charge-coupled device still camera mounted on the 8.2 m Subaru Telescope, makes it possible to probe a wide field (1.77 deg2 field of view) and deep space (26th mag in the i band / epoch). Our primary scientific goals are SN Ia cosmology, Type II supernova (SN II) cosmology, and super-luminous supernova (SLSN) studies, as well as to probe unknown populations of transients. As reported recently (Yasuda et al. 2019), more than 1800 SNe were discovered during a six-month campaign. We deployed machine learning (area under the curve (AUC) boosting) for transient detection (Morii et al. 2016), where the machine determines whether a transient is “real” or “bogus.” In Kimura et al. (2017), we adopted a DNN for SN type classification from a two-dimensional image, and highway block was introduced for the optimization of layers. This research is an extension of our previous work (Kimura et al. 2017) and applies a DNN to the photometric classification of transients. We uniquely attempt to use the observed data in a state that is as “raw” as possible to enable us to directly use the data as input for the machine without fitting the data or extracting characteristics.

The structure of this paper is as follows. We introduce our methods in section 2, and the data in section 3. The design of our DNN model and its application to pseudo-real LSST simulation data is described in section 4. Section 5 presents the application of our model to the actual Subaru/HSC data. We discuss the results in section 6 and conclude the paper in section 7.

2 Methods

2.1 Tasks in HSC survey

The Subaru HSC Transient Survey forms part of the Subaru Strategic Project (SSP), which is a five-year program with a total of 300 dark nights (Aihara et al. 2018a; Miyazaki et al. 2018). The HSC-SSP Transient Survey is composed of two seasons, the first of which was executed for six consecutive months from 2016 November through 2017 April. The HSC is mounted on the prime focus for two weeks in dark time. Weather permitting, we aimed to observe two data points per filter per lunation. Details of the survey strategies and the observing logs were reported in Yasuda et al. (2019), and we used the observed photometric data described there.

One of our primary goals with the HSC-SSP Transient Survey is SN Ia cosmology, which aims to perform the most precise measurement of dark energy at high redshift and test whether the dark energy varies with time (Linder 2003). We have been awarded 96 orbits of Hubble Space Telescope time (WFC3 camera) to execute precise infrared (IR) photometry at the time of maximum. Our HST program uses non-disrupted target of opportunity (ToO), which means we are required to send the request for observation two weeks prior to the observation. In other words, we need to identify good candidates two to three weeks prior to the maximum. Although a high-redshift (⁠|$z$| > 1) time dilation factor helps, it is always a challenge to identify SN Ia on the rise.

Our international collaboration team executes spectroscopic follow-up using large telescopes from around the world. Our target, high-redshift SN Ia, is faint (∼24th mag in the i band) for spectroscopic identification even with the most powerful large telescopes: GMOS Gemini (Hook et al. 2004), GTC OSIRIS,1 Keck LRIS (Oke et al. 1995), VLT FORS (Appenzeller et al. 1998), Subaru FOCAS (Kashikawa et al. 2002), and AAT AAOmega Spectrograph (Saunders et al. 2004). Thus, it is critical to hit the target at the time of maximum brightness either by using ToO (GTC), queue mode (VLT), or classical scheduled observation (Keck, Subaru, ATT).

An SN Ia requires approximately 18 d from its explosion to reach its maximum in the rest frame (Conley et al. 2006; Papadogiannakis et al. 2019). Given this fact, the SNe with which we are concerned are high-redshift SN Ia (⁠|$z$| > 1), in which case we have approximately one month in the observed frame for an SN Ia from the time of explosion until it reaches its maximum. However, our task is to identify these SNe two weeks prior to the maximum, which means we have only two data points per filter. In addition to that, the sky conditions continue to change, and we may not have the data as originally planned. In reality, our identification has to proceed despite data points being missing on the rise.

In parallel to the SN Ia cosmology program, our sister projects also need identification and classification of HSC transients. Specifically, the SN II cosmology program requires timely spectroscopic follow-up to measure the expansion velocity of the photosphere from the Hβ line (de Jaeger et al. 2017). An SLSN is of great interest today, because it is a relatively rare event (Quimby et al. 2011) with diverse origination mechanisms (Gal-Yam 2012; Moriya et al. 2018), and it can be used to probe the high-redshift Universe (Cooke et al. 2012). New types of rapid transients are also discovered by HSC, but their identities are yet to be known (Tampo et al. 2020). For these projects, timely spectroscopic follow-up is also critical (Moriya et al. 2019; Curtin et al. 2019).

An early-phase SN Ia provides us with clues to the explosion mechanism (Maeda et al. 2018) and progenitors (Cao et al. 2015). The advantage of HSC is its ability to survey a large volume, and in practice it has confirmed the long-standing theoretical prediction of helium-shell detonation (Jiang et al. 2017). Finding an early-phase SN Ia is not trivial, but HSC is yielding a new set of early-phase SNe Ia (Jiang et al. 2020). Observations of early-phase core-collapse SNe provide us with crucial information on the sizes of the progenitors (Thompson et al. 2003; Tominaga et al. 2011) and the circumstellar medium (Förster et al. 2018).

2.2 Classification method for the HSC-SSP transient survey

We designed two machine learning models with the emphasis on identifying SN Ia, which requires a time-sensitive trigger for HST IR follow-up. The first model operates in binary mode and classifies whether a transient is an SN Ia. In this regard, the majority of high-redshift transients are known to be of the SN Ia type, and our work entails searching for other unknown transients from among those labeled non-SN Ia. The second model classifies a transient into one of three classes: SN Ia, SN Ibc, or SN II. These three classes were chosen for simplicity, and in fact the majority of SNe belong to one of these three categories. SN Ia is a thermonuclear explosion, and its brightness can be calibrated empirically. SN Ib, SN Ic, and SN II are all core-collapse SNe and are classified by their spectral features (Filippenko 1997). The light curves of SN Ib and SN Ic, which are collectively referred to as SN Ibc, are similar to those of SN Ia and always contaminate the SN Ia cosmology program. They are fainter than SN Ia and redder in general. A major challenge of this work is to determine whether we can distinguish SN Ibc from SN Ia.

3 Data

In this section we present the dataset we used for our study. We first introduce our SN dataset from the HSC-SSP Transient Survey (subsection 3.1). Then we describe the simulated photometric data to train the machine (subsection 3.2). Lastly, we explain the preprocessing of the above data for input into the machine (subsection 3.3).

3.1 Observed data from the Subaru/HSC-SSP transient survey

The goal of this project is to classify the light curves observed by Subaru/HSC. The discovery of 1824 SNe recorded during the six-month HSC-SSP Transient Survey during the period of 2016 November through 2017 April was described in Yasuda et al. (2019). The survey is composed of the Ultra-Deep and Deep COSMOS layers (Scoville et al. 2007). The median 5 σ limiting magnitudes per epoch are 26.4, 26.3, 26.0, 25.6, and 24.6 mag (AB) for the g, r2, i2, |$z$|⁠, and y bands, respectively, for the Ultra-Deep layer. For the Deep layer, the depth is 0.6 mag shallower.

The SN dataset consists of time series photometric data (flux, magnitude, and their errors) in each band for each SN. Because part of the y-band photometric data contains residuals due to improper background subtraction influenced by scattered light (Aihara et al. 2018b), we excluded the y-band data from our study, considering the impact thereof on the classification performance.

The redshift information for HSC SNe is a combination of our follow-up observation results and catalogs from multiple surveys of those host galaxies. The spectral redshifts (spec-|$z$|⁠) are adopted from the results of the follow-up spectrum observations by AAT/AAOmega performed in 2018 and those from the DEIMOS (Hasinger et al. 2018), FMOS-COSMOS (Silverman et al. 2015), C3R2 (Masters et al. 2017), PRIMUS (Coil et al. 2011), and COSMOS catalogs. For those without spec-|$z$|⁠, the photometric redshifts (photo-|$z$|⁠) were adopted from the COSMOS 2015 catalog (Laigle et al. 2016) and those calculated from the HSC-SSP survey data (Tanaka et al. 2018).

3.2 Simulated data for training

We are in need of simulating observed photometric data to train the machine. For normal SN Ia, we used the SALT2 (Guy et al. 2010) model (version 2.4) that requires two input parameters: c for color and x1 for stretch. We adopted an asymmetric Gaussian distribution for c, and x1 from Mosher et al. (2014), and generated light curves and simulated photometric data points based on the observation schedule. Apart from SN Ia, we used published spectral time series from Kessler et al. (2019) which contains both observed core-collapse SN data and light curves from the simulation. We combined an equal number of SN Ia and non-SN Ia observations to approximate the observed fractions. Although these fractions do not need to be exact for the purpose of our classification, it is important to avoid using an unbalanced dataset. For the three-class classification, we set the ratio of SN Ia : SN Ibc : SN II = 10  : 3 : 7. The redshift distribution of galaxies was taken from the COSMOS survey (Laigle et al. 2016), and we distributed the simulated SNe accordingly from |$z$| = 0.1 through |$z$| = 2.0. Throughout the study reported in this paper we used ΛCDM cosmology with Ωm = 0.3 and h = 0.7. We used the filter response including system throughput from Kawanomoto et al. (2018). Examples of simulated photometric data are shown in figure 1.

Overlay plots of simulated light curves with $z$ between 0.1 and 1.2. Each panel shows the plots of SN Ia, Ibc, II data from the left, and the g, r2, i2, $z$, and y bands from the bottom. The variation of the curves in each panel depends on the different parameters and templates used in the simulation. A noise component was not added to these light curves. (Color online)
Fig. 1.

Overlay plots of simulated light curves with |$z$| between 0.1 and 1.2. Each panel shows the plots of SN Ia, Ibc, II data from the left, and the g, r2, i2, |$z$|⁠, and y bands from the bottom. The variation of the curves in each panel depends on the different parameters and templates used in the simulation. A noise component was not added to these light curves. (Color online)

A complication that arises when attempting to simulate realistic data is that the machine does not accept expected errors. Alternatively, we may not have identified a good method for including errors. In this study, we therefore simply used brute force, i.e., we placed the expected photometric error on top of the simulated data such that a simulated data point behaves similarly to one of the many realizations. The magnitude of the error in the HSC the simulation would have to consider varies from night to night because of sky conditions. We measured and derived the flux vs. error relationship from the actual observed data at the simulating epoch and applied that relationship to the simulated photometric data.

Guided by the “accuracy” (subsection 4.2), we determined the number of light curves that would be required for training. Based on our convergence test, we concluded that we would need to generate more than 100000 light curves for training as shown in figure 2. We omitted curves with less than 3 σ detection at the maximum because our detection criterion was 5 σ. For training, we generated 514954 light curves for HSC observed data. Their final class ratio after these criteria is SN Ia : Ibc : II = 0.59 : 0.07 : 0.34, and their peak timings are randomly shifted by 450 d.

Convergence test to determine the number of light curves that would need to be generated to train the machine. The solid line shows the mean accuracy of five classifiers. The shaded area shows the standard deviation of the classifiers, which were trained with five-fold cross validation using the training dataset. The results indicate that more than 100000 light curves would be required for training. (Color online)
Fig. 2.

Convergence test to determine the number of light curves that would need to be generated to train the machine. The solid line shows the mean accuracy of five classifiers. The shaded area shows the standard deviation of the classifiers, which were trained with five-fold cross validation using the training dataset. The results indicate that more than 100000 light curves would be required for training. (Color online)

3.3 Preprocessing of input data

Based on our pre-experiment with the simulated dataset, we found the machine to perform best by using a combination of the normalized flux (f) and pseudo-absolute magnitude (M):
(1)
where |$f_{i}^{\mathrm{scale}}$| is the ith raw observed flux normalized by its maximum flux:
(2)
and |$M_i^\mathrm{abs}$| is the ith pseudo-absolute observed magnitude. For simplicity, we ignored K-correction and used the distance modulus [DM(z)] based on ΛCDM with the photometric redshift from its host galaxy:
(3)
We can justify this operation because the training set and the observed dataset are processed using the same approach. In the case of the existence of a K-correction offset, both datasets would experience this in the same way. In addition, because the observed flux could take on a negative value owing to statistical fluctuation, we adopt hyperbolic sine to imitate the magnitude system we use (Lupton et al. 1999):
(4)
In fact, combination of the flux and magnitudes is redundant, because knowledge of one would enable us to calculate the other explicitly. However, based on our experiment, the score of the machine improves by using both. We suspect that the distribution in flux (linear) differs from that of the magnitude space (log), which provides the machine with additional information. Thus, we used the pseudo-absolute magnitude (M) and normalized flux (f) as input.

4 Deep neural network classifier

With the rise of Big Data, the use of machine learning techniques has played a critical role in the analysis of astronomical data. Techniques such as random forest, support vector machine, and convolution neural network have been used for photometric data analysis (Pasquet et al. 2019), galaxy classifications (Hausen & Robertson 2020), and spectral classifications (Garcia-Dias et al. 2018; Muthukrishna et al. 2019b; Sharma et al. 2020).

In our work, we seek to classify SNe from photometric data. Our approach entails making use of the observed data without preprocessing or parameterization. In this regard, we rely on deep learning to make our work possible. We decided to test the extent to which deep learning could provide useful results without extracting features such as color, the width of the light curve, and the peak magnitude.

The fact that we went one step further by leaving the observed data as raw as possible means that our input consists of a simple array of magnitudes. An attempt such as this would not have been possible ten years ago; however, owing to advancements in computing and the deep learning technique, this has become a reality. Among the many machine learning methods, we decided to use a DNN to enable us to classify astronomical objects from the raw observed data.

4.1 Model design

In this section we describe the design of our DNN model,2 which accepts an array of observed magnitudes as its input and outputs the SN classification with probabilities. We adopted a highway layer (also known as a “layer in layer”: Srivastava et al. 2015a) as a core part of our network. Compared to plain DNN, the performance of a highway layer improves when the network is deep in terms of parameter optimization (Srivastava et al. 2015b).

Similar to other DNN models, this model proceeds through a training and validation process to optimize the parameters, and we describe the steps below. Our terminology is commonly used in the world of DNN, but as this is a new introduction to the astronomical community, we explain each step in detail. The architecture of our model is summarized in figure 3. Ultimately, each SN is assigned a probability of belonging to a certain astrophysical class, in our case, the type of SN.

Architecture of the deep neural network classifier. The green boxes are parameters optimized by the gradient descent method during training. The red boxes are hyperparameters that are optimized during the hyperparameter search. The batch normalization layer has four variables (μ, σ2, γ, and β), where μ and σ2 are intended to learn the statistics (mean and variance) of the value through the layer, respectively, and γ and β are scale and shift parameters, respectively, to adjust the output. Note that μ and σ2 are not updated by gradient descent; instead, they are updated by the moving average. They were omitted from the figure for simplicity. (Color online)
Fig. 3.

Architecture of the deep neural network classifier. The green boxes are parameters optimized by the gradient descent method during training. The red boxes are hyperparameters that are optimized during the hyperparameter search. The batch normalization layer has four variables (μ, σ2, γ, and β), where μ and σ2 are intended to learn the statistics (mean and variance) of the value through the layer, respectively, and γ and β are scale and shift parameters, respectively, to adjust the output. Note that μ and σ2 are not updated by gradient descent; instead, they are updated by the moving average. They were omitted from the figure for simplicity. (Color online)

4.1.1 Input

Our input is an array of magnitudes and normalized fluxes of the ith SN in the training dataset:
(5)
We do not explicitly specify the time at which the data were recorded or the particular filter used, but this information is recorded as an order inside the array. The philosophy here is that the training set, which is composed of simulated data of the same array length, holds information on the filter and dates. For example, the jth magnitude in the array is data recorded on a certain date and by a certain filter. The combination of the date and filter is identical to those in the training set. Therefore, the jth component implicitly contains unique information about the dates and filters. Considering that the input consists of a combination of the magnitude and normalized flux, the size of our input array is 1 × 2N per SN, where N is the number of data points.

4.1.2 First fully connected layer

We decided to make use of D neurons, also known as the number of “hidden layers,” and D is greater than the number of input components (2N). However, the dimension of D is not known in advance, and this is one of the hyperparameters we optimize later in this section. Because the dimensionality of the input (2N) could differ from the number of optimized neurons D, we would need to adjust the number of dimensions and that is the role of this first fully connected layer F(x),
(6)
F(x) is given by a linear combination of a matrix |$W_{\rm fc} \in \mathbb {R}^{D\times 2N}$| and a vector |$b_{\rm fc} \in \mathbb {R}^D$|⁠. The initial value of Wfc is generated by a Gaussian distribution and bfc is initialized to |$\mathbf {0}$|⁠, which is a D-dimensional vector of which all the elements are zero. We used the Python wrapper library dm-sonnet (version 1.23)3 and its function linear to supply the F(x) when plugging in “2N” and “D.” Subsequently, Wfc and bfc are optimized by the open source machine learning package Tensorflow (version 1.14) (Abadi et al. 2016). Unless stated otherwise, we used the libraries from Tensorflow.

4.1.3 Dropout layer

To obtain a robust result, it is always best to train all of the neurons as an ensemble and avoid a situation in which one of the neurons adversely affects the result. Dropout is a process in which certain neurons are randomly dropped from training and the dropout rate can be optimized as one of the hyperparameters (Srivastava et al. 2014).

4.1.4 Highway layer

We adopted a highway layer (Srivastava et al. 2015a) and optimized the number of layers therein during the hyperparameter search. In theory, it would be possible to design a very deep layer with more layers than would be necessary. However, in reality, it is not trivial to optimize the number of layers. The depth and/or size of the layers of the DNN model are directly related to the complexity of the features used as input, and greatly affect the computational performance of the task. Thus, an overly deep model would complicate the learning process and cause performance degradation. A highway layer is a technique that stabilizes the learning process by devising the network structure. We previously used a highway layer, which we tested on two-dimensional images, and it delivered good performance (Kimura et al. 2017). Details of the use and advantages of the highway layer are provided in Kimura et al. (2017). This encouraged us to adopt a highway layer scheme for this analysis. The output of the highway layer is calculated from the values of multiple paths. The output, |$\mathbf {Highway}({x})$|⁠, is formulated as
(7)
where H is a nonlinear transformation layer, G is the transformation gate function layer and controls the transformation of the input, C is the carry gate function layer, and ⊗ provides the element-wise product, also known as the Hadamard product. A highway layer includes several other layers, a structure known as “layer in layer.” Each function is defined as follows:
(8)
(9)
(10)
where a is an activation function, namely, sigmoid,
(11)
(12)
where D is the number of neurons. Each element of G(x) always takes a value between 0 and 1. Eventually, the Highway layer behaves as follows:
(13)
Along with the dimensions of the hidden layer D, the dropout ratio, batch normalization, and the types of activation function, the number of repetitions T is one of the hyperparameters and is optimized by performing a hyperparameter search. Details are provided in subsection 4.2.

4.1.5 Batch normalization layer

We adopt batch normalization (Ioffe & Szegedy 2015) to accelerate and stabilize the optimization. Even if a large number of parameters need to be trained, batch normalization facilitates convergence of the training process, reduces errors on the slope when we apply entropy minimization, prevents the average and dispersion from becoming exponentially large in deep layers, and minimizes the biases on outputs (Bjorck et al. 2018). Batch normalization is effective in many cases. However, the performance of a model that employs both batch normalization and dropout may degrade (Li et al. 2019).

4.1.6 Activation layer

Each neuron is activated by using nonlinear transformation. Nonlinearity is an important component of DNN, because it allows a wide variety of expressions. Note that the majority of the layers, including fully connected layers, involve a linear transformation and, even if a number of layers were to exist, it would be equivalent to one single linear transformation. Thus, nonlinear transformation is essential to allow each neuron the freedom to have any necessary values. For the first iteration, we do not know what kind of transformation is the best; thus, the transformation itself is taken as one of the hyperparameters. In our work, we used the functions “tf.nn” in Tensorflow.

4.1.7 Second fully connected layer

After T repetitions of the highway layer, batch normalization layer, and activation layer, it is necessary to convert the number of neurons to the number of SNe times the number of SN classes. This operation is opposite to that of the first fully connected layer.

4.1.8 Softmax layer

The softmax layer normalizes the input value of this layer, which is denoted by |$h \in \mathbb {R}^D$|⁠. The output value is |$\hat{y} \in \mathbb {R}^D$|⁠, and each element |$\hat{y}_k$| is expressed as
(14)
The normalized value |$\hat{y}$| satisfies |$\hat{y}_k \ge 0$| and |$\sum _k^K \hat{y}_k =1$|⁠. We can interpret |$\hat{y}_k$| as the probability that the input x belongs to class k. However, we note that this is a pseudo-probability and that it differs from the statistical probability.

4.2 Hyperparameter search

We perform a hyperparameter search by combining grid search and the tree-structured Parzen estimator (TPE) algorithm (Bergstra et al. 2013). Although grid search is not suitable to search a high-dimensional space, it has the advantage of searching for multiple points in parallel. In addition, because of the simplicity of the algorithm, it allows us to convey knowledge acquired during preliminary experiments for parameter initialization. Meanwhile, the TPE algorithm is suitable for searching a high-dimensional space, but has the disadvantage of not knowing where to start initially. Therefore, this time, our search was guided by the hyperparameter values that were obtained in the preliminary experiment using grid search, and these results were then used as input for the TPE algorithm. The ranges in which we searched for the hyperparameters are given in table 1.

Table 1.

Ranges of hyperparameter search for type classification.

HyperparameterValue (grid)Range (TPE)
D{100, 300}50, …, 1000
T{1, 3, 5}1, …, 5
bn{true}{true, false}
Drop rate[5 × 10−3, 0.035][5 × 10−4, 0.25]
Type{identity, relu, sigmoid, tanh}
HyperparameterValue (grid)Range (TPE)
D{100, 300}50, …, 1000
T{1, 3, 5}1, …, 5
bn{true}{true, false}
Drop rate[5 × 10−3, 0.035][5 × 10−4, 0.25]
Type{identity, relu, sigmoid, tanh}
Table 1.

Ranges of hyperparameter search for type classification.

HyperparameterValue (grid)Range (TPE)
D{100, 300}50, …, 1000
T{1, 3, 5}1, …, 5
bn{true}{true, false}
Drop rate[5 × 10−3, 0.035][5 × 10−4, 0.25]
Type{identity, relu, sigmoid, tanh}
HyperparameterValue (grid)Range (TPE)
D{100, 300}50, …, 1000
T{1, 3, 5}1, …, 5
bn{true}{true, false}
Drop rate[5 × 10−3, 0.035][5 × 10−4, 0.25]
Type{identity, relu, sigmoid, tanh}

According to the usual approach, we divided the dataset into training and validation datasets. We used the training data to optimize the DNN, and the validation data to measure the accuracy. The hyperparameters were optimized by evaluation with the validation dataset to maximize the accuracy of this dataset. This process was iteratively conducted 100 times to allow the accuracy to converge to its maximum (figure 4).

Result of iterative hyperparameter search (100 cycles) showing its convergence to its maximum performance in terms of accuracy. The task involved binary classification. (Color online)
Fig. 4.

Result of iterative hyperparameter search (100 cycles) showing its convergence to its maximum performance in terms of accuracy. The task involved binary classification. (Color online)

We can train the DNN classifier in the same way regardless of the number of classes. In the case of multi-type classification, the number of classes is K = 3 in our experiment; thus, the number of outputs of the DNN classifier is also three. In binary classification (SN Ia or non-SN Ia), the number of outputs is two.

We trained the model by minimizing the cross-entropy error:
(15)
where y is the ground truth vector, which entails one-hot encoding of K dimensions, and |$\hat{y}$| is the DNN output vector. We deployed the Adam optimizer (Kingma & Ba 2014) which uses a stochastic gradient method to optimize the model parameters.

We introduced data augmentation to prevent overfitting at the time of training. By increasing the number of input data using data augmentation, we prevent DNN from having to memorize the entire training dataset. We used two data augmentation methods to augment the training dataset. The first was to add Gaussian noise (based on the expected observed uncertainty) to the simulated flux. The second involved the use of the mixup technique (Zhang et al. 2017).

Mixup generates a new virtual training dataset as follows:
(16)
(17)
where (xu, yu) and (x|$v$|, y|$v$|) are samples drawn at random from the training dataset, x is the input vector, y is the one-hot label vector, and the mixing ratio λ ∈ [0, 1] is drawn from a random distribution, of which the density is low near 0 and 1 and higher near 0.5. The datasets generated in this way are suitable to enable the DNN to learn the classification boundaries.

As described above, the hyperparameters (the red boxes in figure 3) of the model are optimized by maximizing the accuracy, whereas the model parameters (the green boxes in figure 3) are optimized by minimizing the cross-entropy error.

4.3 Testing the DNN model with the PLAsTiCC dataset

Before we applied our model to the data observed by HSC, we tested it with a dataset resulting from the LSST simulated classification challenge, i.e., the PLAsTiCC dataset (The PLAsTiCC team 2018; Malz et al. 2019), which is composed of realistic photometric data with errors on time-variable objects. To evaluate our model, we required a dataset with labels of true identity. The PLAsTiCC Deep Drilling Field (DDF) dataset contains data similar to those in the HSC-SSP Transient Survey, and we took advantage thereof. However, we generated the training set ourselves and selected not to use the training set provided by the PLAsTiCC team because we knew that the size of their training dataset was insufficient to achieve maximum performance (figure 2).

The training dataset was created using the method described in subsection 3.2. We generated 370345 light curves based on the filter response and photometric zero-point for LSST (Ivezić et al. 2019). These light curves were composed of the different types of SNe in the ratio SN Ia : ratio Ibc : II = 0.60 : 0.06 : 0.34, with their peaks randomly shifted in time. The test dataset was created by extracting 2297 light curves from the PLAsTiCC dataset. These light curves were labeled Ia, Ibc, or II, to identify the type of SN each curve represented. The light curves were simulated to occur in the COSMOS field.

We used the AUC of the receiver operating characteristic (ROC) curve, the precision-recall curve in two-class classification, and the accuracy from the confusion matrix in three-class classification as metrics to evaluate our model. Different combinations of inputs were tested to determine which performs best when using the PLAsTiCC dataset. Our input could be a combination of the arrays of normalized flux (f), magnitude (m), or pseudo-absolute magnitude (M). Table 2 lists the AUC for two-class classification and the accuracy for three-class classification when using PLAsTiCC data. Our investigation showed that a combination of the normalized flux (f) and pseudo-absolute magnitude (M) performs best, and, although the information is redundant, we suspect the different distribution of the data provides the machine with additional guidance. The AUC values for the ROC curve and the precision-recall curve are 0.996 and 0.995, respectively. Figure 5 shows the confusion matrix for the three-class classification, with the total accuracy calculated as 95.3%. As is always the case in the real world, it is difficult to classify SN Ibc, but the effect on overall accuracy is relatively small.

Normalized confusion matrix in the three-class classification of the PLAsTiCC dataset. The classifier received the pseudo-absolute magnitude and normalized flux as its input. The proportions in each row sum to 1. The numbers in parentheses represent the raw numbers. (Color online)
Fig. 5.

Normalized confusion matrix in the three-class classification of the PLAsTiCC dataset. The classifier received the pseudo-absolute magnitude and normalized flux as its input. The proportions in each row sum to 1. The numbers in parentheses represent the raw numbers. (Color online)

Table 2.

Classification performance of each input for the PLAsTiCC dataset.

Input*AUCAccuracy
MmfROCPre.-rec.
|$\checkmark$||$\checkmark$|0.9960.9950.953
|$\checkmark$|0.9950.9930.952
|$\checkmark$||$\checkmark$|0.9950.9930.948
|$\checkmark$|0.9950.9910.940
Input*AUCAccuracy
MmfROCPre.-rec.
|$\checkmark$||$\checkmark$|0.9960.9950.953
|$\checkmark$|0.9950.9930.952
|$\checkmark$||$\checkmark$|0.9950.9930.948
|$\checkmark$|0.9950.9910.940
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

Table 2.

Classification performance of each input for the PLAsTiCC dataset.

Input*AUCAccuracy
MmfROCPre.-rec.
|$\checkmark$||$\checkmark$|0.9960.9950.953
|$\checkmark$|0.9950.9930.952
|$\checkmark$||$\checkmark$|0.9950.9930.948
|$\checkmark$|0.9950.9910.940
Input*AUCAccuracy
MmfROCPre.-rec.
|$\checkmark$||$\checkmark$|0.9960.9950.953
|$\checkmark$|0.9950.9930.952
|$\checkmark$||$\checkmark$|0.9950.9930.948
|$\checkmark$|0.9950.9910.940
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

Table 3 summarizes the classification performance for each group of the test set divided according to the maximum signal-to-noise ratio (SNR) of the photometric data. It shows that the classification performance tends to improve as the maximum signal-to-noise ratio increases.

Table 3.

Classification performance for the maximum signal-to-noise ratio.

Max. SNRNumberAUC*
ROCPre.-rec.
<5310.9750.983
5–107360.9890.983
10–208140.9990.998
>207160.9990.999
All22970.9960.995
Max. SNRNumberAUC*
ROCPre.-rec.
<5310.9750.983
5–107360.9890.983
10–208140.9990.998
>207160.9990.999
All22970.9960.995
*

AUC for the best performing model (M + f).

Table 3.

Classification performance for the maximum signal-to-noise ratio.

Max. SNRNumberAUC*
ROCPre.-rec.
<5310.9750.983
5–107360.9890.983
10–208140.9990.998
>207160.9990.999
All22970.9960.995
Max. SNRNumberAUC*
ROCPre.-rec.
<5310.9750.983
5–107360.9890.983
10–208140.9990.998
>207160.9990.999
All22970.9960.995
*

AUC for the best performing model (M + f).

In the three-class classification of the PLAsTiCC dataset, 107 SNe were misclassified and have the following characteristics:

  • 54% (58/107) of them were “incomplete events” that did not include the peak phase of the SN (the period of 10 d before and 20 d after the peak in the observed frame) in the photometric data, whereas they only constitute 38% of all events.

  • Of the remaining misclassifications, 29% (14/49) are on the boundary where the difference in probability between the correct class and the predicted class is less than 0.1.

  • In more than half of the remaining 35 events, SN Ibc was misclassified as either SN Ia or II.

Figure 6 shows the accuracy against redshift for the PLAsTiCC dataset. The accuracy for SN Ibc is lower than that of the other classes; in particular, it is greatly reduced at redshifts beyond 1.0, and also decreased at redshifts of 0.1 to 0.2. Manual verification of individual misclassifications revealed that, although certain misclassified SNe are too faint to classify even with conventional methods, a few bright SNe were also completely misclassified.

Accuracy as a function of redshift in the three-class classification of the PLAsTiCC dataset. (Color online)
Fig. 6.

Accuracy as a function of redshift in the three-class classification of the PLAsTiCC dataset. (Color online)

5 Application to the HSC-SSP transient survey

We applied the developed classifier to the dataset acquired during the HSC-SSP Transient Survey. This dataset includes photometric data of 1824 SNe. As described in Yasuda et al. (2019), the survey was conducted in two layers with different depths and cadence, i.e., “Deep” and “Ultra-Deep.” Therefore, the number of photometric data points of an SN in each layer could be different. Our DNN model requires exactly the same number of data points as its input; thus, we divided our dataset into five cases based on the number of photometric data points. The number of SNe for each case is summarized in table 4. For example, the number of SNe in Case 0 is 709, and they are in the Ultra-Deep layer. Each SN is represented by a total of 42 epochs of photometric data in four bands (g, r2, i2, and |$z$|⁠). The number of epochs and filter schedule for Case 0 SNe are summarized in table 5. The introduction of these five cases enabled the machine to correctly classify 1812 HSC SNe, which corresponds to 99.3% of the 1824 SNe. The remaining 12 SNe were excluded owing to missing data.

Table 4.

Number of SNe for each case.

CaseEpochNumberFraction
0427090.391
1266460.357
2192710.150
3101220.067
44640.035
CaseEpochNumberFraction
0427090.391
1266460.357
2192710.150
3101220.067
44640.035
Table 4.

Number of SNe for each case.

CaseEpochNumberFraction
0427090.391
1266460.357
2192710.150
3101220.067
44640.035
CaseEpochNumberFraction
0427090.391
1266460.357
2192710.150
3101220.067
44640.035
Table 5.

Number of input epochs and the schedule for Case 0 SNe.

FilterEpochsElapsed day
g82, 40, 63, 70, 92, 119, 126, 154
r295, 32, 61, 71, 92, 103, 122, 129, 151
i2132, 6, 32, 40, 61, 68, 71, 94, 101, 120, 127, 154, 155
|$z$|120, 6, 30, 40, 59, 68, 90, 101, 119, 126, 151, 157
FilterEpochsElapsed day
g82, 40, 63, 70, 92, 119, 126, 154
r295, 32, 61, 71, 92, 103, 122, 129, 151
i2132, 6, 32, 40, 61, 68, 71, 94, 101, 120, 127, 154, 155
|$z$|120, 6, 30, 40, 59, 68, 90, 101, 119, 126, 151, 157
Table 5.

Number of input epochs and the schedule for Case 0 SNe.

FilterEpochsElapsed day
g82, 40, 63, 70, 92, 119, 126, 154
r295, 32, 61, 71, 92, 103, 122, 129, 151
i2132, 6, 32, 40, 61, 68, 71, 94, 101, 120, 127, 154, 155
|$z$|120, 6, 30, 40, 59, 68, 90, 101, 119, 126, 151, 157
FilterEpochsElapsed day
g82, 40, 63, 70, 92, 119, 126, 154
r295, 32, 61, 71, 92, 103, 122, 129, 151
i2132, 6, 32, 40, 61, 68, 71, 94, 101, 120, 127, 154, 155
|$z$|120, 6, 30, 40, 59, 68, 90, 101, 119, 126, 151, 157

We subjected the aforementioned five cases of observed HSC data to both two-class and three-class classification. For each of these cases, the machine needs to be trained independently with a dedicated training dataset. Thus, the hyperparameters were optimized for each case and are reported in table 6. Subsections 5.1 and 5.2 describe the performance evaluation for each classification.

Table 6.

Optimized hyperparameters for classification.

Input*Two-classThree-class
MmfTDDrop ratebnTypeTDDrop ratebnType
Case 0|$\checkmark$||$\checkmark$|51789.47 × 10−31sigmoid44291.20 × 10−30linear
|$\checkmark$|32479.68 × 10−41sigmoid45162.54 × 10−30tanh
|$\checkmark$||$\checkmark$|45316.43 × 10−30linear46081.72 × 10−20linear
|$\checkmark$|44119.00 × 10−21sigmoid48381.36 × 10−30tanh
Case 1|$\checkmark$||$\checkmark$|57348.75 × 10−40tanh49151.03 × 10−20linear
|$\checkmark$|23897.17 × 10−21sigmoid56982.79 × 10−20linear
|$\checkmark$||$\checkmark$|26477.92 × 10−40tanh45401.45 × 10−30linear
|$\checkmark$|23421.42 × 10−21sigmoid25209.99 × 10−41sigmoid
Case 2|$\checkmark$||$\checkmark$|23681.79 × 10−31sigmoid46989.08 × 10−40linear
|$\checkmark$|49201.44 × 10−30sigmoid46147.03 × 10−30linear
|$\checkmark$||$\checkmark$|55724.27 × 10−31sigmoid48968.58 × 10−30linear
|$\checkmark$|46401.12 × 10−11sigmoid53005.00 × 10−11sigmoid
Case 3|$\checkmark$||$\checkmark$|58931.42 × 10−31sigmoid45225.02 × 10−40tanh
|$\checkmark$|48801.98 × 10−21sigmoid58414.96 × 10−21sigmoid
|$\checkmark$||$\checkmark$|33005.00 × 10−31linear34629.28 × 10−41sigmoid
|$\checkmark$|39309.77 × 10−21sigmoid33005.00 × 10−31sigmoid
Case 4|$\checkmark$||$\checkmark$|53794.21 × 10−31sigmoid34842.13 × 10−30linear
|$\checkmark$|26311.77 × 10−20sigmoid42433.21 × 10−31sigmoid
|$\checkmark$||$\checkmark$|51404.04 × 10−31sigmoid53891.23 × 10−40tanh
|$\checkmark$|45675.77 × 10−20sigmoid33542.14 × 10−11sigmoid
Input*Two-classThree-class
MmfTDDrop ratebnTypeTDDrop ratebnType
Case 0|$\checkmark$||$\checkmark$|51789.47 × 10−31sigmoid44291.20 × 10−30linear
|$\checkmark$|32479.68 × 10−41sigmoid45162.54 × 10−30tanh
|$\checkmark$||$\checkmark$|45316.43 × 10−30linear46081.72 × 10−20linear
|$\checkmark$|44119.00 × 10−21sigmoid48381.36 × 10−30tanh
Case 1|$\checkmark$||$\checkmark$|57348.75 × 10−40tanh49151.03 × 10−20linear
|$\checkmark$|23897.17 × 10−21sigmoid56982.79 × 10−20linear
|$\checkmark$||$\checkmark$|26477.92 × 10−40tanh45401.45 × 10−30linear
|$\checkmark$|23421.42 × 10−21sigmoid25209.99 × 10−41sigmoid
Case 2|$\checkmark$||$\checkmark$|23681.79 × 10−31sigmoid46989.08 × 10−40linear
|$\checkmark$|49201.44 × 10−30sigmoid46147.03 × 10−30linear
|$\checkmark$||$\checkmark$|55724.27 × 10−31sigmoid48968.58 × 10−30linear
|$\checkmark$|46401.12 × 10−11sigmoid53005.00 × 10−11sigmoid
Case 3|$\checkmark$||$\checkmark$|58931.42 × 10−31sigmoid45225.02 × 10−40tanh
|$\checkmark$|48801.98 × 10−21sigmoid58414.96 × 10−21sigmoid
|$\checkmark$||$\checkmark$|33005.00 × 10−31linear34629.28 × 10−41sigmoid
|$\checkmark$|39309.77 × 10−21sigmoid33005.00 × 10−31sigmoid
Case 4|$\checkmark$||$\checkmark$|53794.21 × 10−31sigmoid34842.13 × 10−30linear
|$\checkmark$|26311.77 × 10−20sigmoid42433.21 × 10−31sigmoid
|$\checkmark$||$\checkmark$|51404.04 × 10−31sigmoid53891.23 × 10−40tanh
|$\checkmark$|45675.77 × 10−20sigmoid33542.14 × 10−11sigmoid
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

Table 6.

Optimized hyperparameters for classification.

Input*Two-classThree-class
MmfTDDrop ratebnTypeTDDrop ratebnType
Case 0|$\checkmark$||$\checkmark$|51789.47 × 10−31sigmoid44291.20 × 10−30linear
|$\checkmark$|32479.68 × 10−41sigmoid45162.54 × 10−30tanh
|$\checkmark$||$\checkmark$|45316.43 × 10−30linear46081.72 × 10−20linear
|$\checkmark$|44119.00 × 10−21sigmoid48381.36 × 10−30tanh
Case 1|$\checkmark$||$\checkmark$|57348.75 × 10−40tanh49151.03 × 10−20linear
|$\checkmark$|23897.17 × 10−21sigmoid56982.79 × 10−20linear
|$\checkmark$||$\checkmark$|26477.92 × 10−40tanh45401.45 × 10−30linear
|$\checkmark$|23421.42 × 10−21sigmoid25209.99 × 10−41sigmoid
Case 2|$\checkmark$||$\checkmark$|23681.79 × 10−31sigmoid46989.08 × 10−40linear
|$\checkmark$|49201.44 × 10−30sigmoid46147.03 × 10−30linear
|$\checkmark$||$\checkmark$|55724.27 × 10−31sigmoid48968.58 × 10−30linear
|$\checkmark$|46401.12 × 10−11sigmoid53005.00 × 10−11sigmoid
Case 3|$\checkmark$||$\checkmark$|58931.42 × 10−31sigmoid45225.02 × 10−40tanh
|$\checkmark$|48801.98 × 10−21sigmoid58414.96 × 10−21sigmoid
|$\checkmark$||$\checkmark$|33005.00 × 10−31linear34629.28 × 10−41sigmoid
|$\checkmark$|39309.77 × 10−21sigmoid33005.00 × 10−31sigmoid
Case 4|$\checkmark$||$\checkmark$|53794.21 × 10−31sigmoid34842.13 × 10−30linear
|$\checkmark$|26311.77 × 10−20sigmoid42433.21 × 10−31sigmoid
|$\checkmark$||$\checkmark$|51404.04 × 10−31sigmoid53891.23 × 10−40tanh
|$\checkmark$|45675.77 × 10−20sigmoid33542.14 × 10−11sigmoid
Input*Two-classThree-class
MmfTDDrop ratebnTypeTDDrop ratebnType
Case 0|$\checkmark$||$\checkmark$|51789.47 × 10−31sigmoid44291.20 × 10−30linear
|$\checkmark$|32479.68 × 10−41sigmoid45162.54 × 10−30tanh
|$\checkmark$||$\checkmark$|45316.43 × 10−30linear46081.72 × 10−20linear
|$\checkmark$|44119.00 × 10−21sigmoid48381.36 × 10−30tanh
Case 1|$\checkmark$||$\checkmark$|57348.75 × 10−40tanh49151.03 × 10−20linear
|$\checkmark$|23897.17 × 10−21sigmoid56982.79 × 10−20linear
|$\checkmark$||$\checkmark$|26477.92 × 10−40tanh45401.45 × 10−30linear
|$\checkmark$|23421.42 × 10−21sigmoid25209.99 × 10−41sigmoid
Case 2|$\checkmark$||$\checkmark$|23681.79 × 10−31sigmoid46989.08 × 10−40linear
|$\checkmark$|49201.44 × 10−30sigmoid46147.03 × 10−30linear
|$\checkmark$||$\checkmark$|55724.27 × 10−31sigmoid48968.58 × 10−30linear
|$\checkmark$|46401.12 × 10−11sigmoid53005.00 × 10−11sigmoid
Case 3|$\checkmark$||$\checkmark$|58931.42 × 10−31sigmoid45225.02 × 10−40tanh
|$\checkmark$|48801.98 × 10−21sigmoid58414.96 × 10−21sigmoid
|$\checkmark$||$\checkmark$|33005.00 × 10−31linear34629.28 × 10−41sigmoid
|$\checkmark$|39309.77 × 10−21sigmoid33005.00 × 10−31sigmoid
Case 4|$\checkmark$||$\checkmark$|53794.21 × 10−31sigmoid34842.13 × 10−30linear
|$\checkmark$|26311.77 × 10−20sigmoid42433.21 × 10−31sigmoid
|$\checkmark$||$\checkmark$|51404.04 × 10−31sigmoid53891.23 × 10−40tanh
|$\checkmark$|45675.77 × 10−20sigmoid33542.14 × 10−11sigmoid
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

5.1 Binary classification

Binary classification was performed using four versions of classifiers, which we prepared with different inputs as in the PLAsTiCC dataset. This approach allowed us to compare their performance, a summary of which is provided in table 7.

Table 7.

AUC of each input in the HSC binary classification.

DatasetInput*ROCPrecision-recall
MmfCase 01234AllCase 01234All
Validation|$\checkmark$||$\checkmark$|1.0000.9900.9870.9760.8870.9931.0000.9930.9910.9830.9170.995
|$\checkmark$|0.9990.9800.9750.9590.8450.9870.9990.9860.9820.9720.8860.991
|$\checkmark$||$\checkmark$|0.9990.9830.9790.9630.8170.9880.9990.9880.9850.9730.8630.992
|$\checkmark$|0.9960.9710.9660.9380.7900.9800.9970.9800.9760.9550.8410.985
Test|$\checkmark$||$\checkmark$|0.9750.9780.9310.8441.0000.9660.9310.9550.7730.7611.0000.909
(Light curve|$\checkmark$|0.9860.9670.9420.9670.9640.9710.9650.9350.8490.9660.9440.934
verified)|$\checkmark$||$\checkmark$|0.9470.9260.8870.7561.0000.9230.8360.8600.8030.6321.0000.820
|$\checkmark$|0.9450.8640.8540.7560.6430.8960.8260.7690.7520.6120.6870.787
Test|$\checkmark$||$\checkmark$|0.9450.9220.9140.8630.8440.9250.8550.8400.8090.7880.6500.832
(All labeled)|$\checkmark$|0.9570.9090.8790.9080.8640.9220.9010.8140.7140.8850.7020.817
|$\checkmark$||$\checkmark$|0.9150.8890.8850.7180.7110.8850.7800.7780.7680.5430.3630.749
|$\checkmark$|0.9110.8370.8550.7130.6560.8620.7730.6850.7120.5230.3850.712
DatasetInput*ROCPrecision-recall
MmfCase 01234AllCase 01234All
Validation|$\checkmark$||$\checkmark$|1.0000.9900.9870.9760.8870.9931.0000.9930.9910.9830.9170.995
|$\checkmark$|0.9990.9800.9750.9590.8450.9870.9990.9860.9820.9720.8860.991
|$\checkmark$||$\checkmark$|0.9990.9830.9790.9630.8170.9880.9990.9880.9850.9730.8630.992
|$\checkmark$|0.9960.9710.9660.9380.7900.9800.9970.9800.9760.9550.8410.985
Test|$\checkmark$||$\checkmark$|0.9750.9780.9310.8441.0000.9660.9310.9550.7730.7611.0000.909
(Light curve|$\checkmark$|0.9860.9670.9420.9670.9640.9710.9650.9350.8490.9660.9440.934
verified)|$\checkmark$||$\checkmark$|0.9470.9260.8870.7561.0000.9230.8360.8600.8030.6321.0000.820
|$\checkmark$|0.9450.8640.8540.7560.6430.8960.8260.7690.7520.6120.6870.787
Test|$\checkmark$||$\checkmark$|0.9450.9220.9140.8630.8440.9250.8550.8400.8090.7880.6500.832
(All labeled)|$\checkmark$|0.9570.9090.8790.9080.8640.9220.9010.8140.7140.8850.7020.817
|$\checkmark$||$\checkmark$|0.9150.8890.8850.7180.7110.8850.7800.7780.7680.5430.3630.749
|$\checkmark$|0.9110.8370.8550.7130.6560.8620.7730.6850.7120.5230.3850.712
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

Table 7.

AUC of each input in the HSC binary classification.

DatasetInput*ROCPrecision-recall
MmfCase 01234AllCase 01234All
Validation|$\checkmark$||$\checkmark$|1.0000.9900.9870.9760.8870.9931.0000.9930.9910.9830.9170.995
|$\checkmark$|0.9990.9800.9750.9590.8450.9870.9990.9860.9820.9720.8860.991
|$\checkmark$||$\checkmark$|0.9990.9830.9790.9630.8170.9880.9990.9880.9850.9730.8630.992
|$\checkmark$|0.9960.9710.9660.9380.7900.9800.9970.9800.9760.9550.8410.985
Test|$\checkmark$||$\checkmark$|0.9750.9780.9310.8441.0000.9660.9310.9550.7730.7611.0000.909
(Light curve|$\checkmark$|0.9860.9670.9420.9670.9640.9710.9650.9350.8490.9660.9440.934
verified)|$\checkmark$||$\checkmark$|0.9470.9260.8870.7561.0000.9230.8360.8600.8030.6321.0000.820
|$\checkmark$|0.9450.8640.8540.7560.6430.8960.8260.7690.7520.6120.6870.787
Test|$\checkmark$||$\checkmark$|0.9450.9220.9140.8630.8440.9250.8550.8400.8090.7880.6500.832
(All labeled)|$\checkmark$|0.9570.9090.8790.9080.8640.9220.9010.8140.7140.8850.7020.817
|$\checkmark$||$\checkmark$|0.9150.8890.8850.7180.7110.8850.7800.7780.7680.5430.3630.749
|$\checkmark$|0.9110.8370.8550.7130.6560.8620.7730.6850.7120.5230.3850.712
DatasetInput*ROCPrecision-recall
MmfCase 01234AllCase 01234All
Validation|$\checkmark$||$\checkmark$|1.0000.9900.9870.9760.8870.9931.0000.9930.9910.9830.9170.995
|$\checkmark$|0.9990.9800.9750.9590.8450.9870.9990.9860.9820.9720.8860.991
|$\checkmark$||$\checkmark$|0.9990.9830.9790.9630.8170.9880.9990.9880.9850.9730.8630.992
|$\checkmark$|0.9960.9710.9660.9380.7900.9800.9970.9800.9760.9550.8410.985
Test|$\checkmark$||$\checkmark$|0.9750.9780.9310.8441.0000.9660.9310.9550.7730.7611.0000.909
(Light curve|$\checkmark$|0.9860.9670.9420.9670.9640.9710.9650.9350.8490.9660.9440.934
verified)|$\checkmark$||$\checkmark$|0.9470.9260.8870.7561.0000.9230.8360.8600.8030.6321.0000.820
|$\checkmark$|0.9450.8640.8540.7560.6430.8960.8260.7690.7520.6120.6870.787
Test|$\checkmark$||$\checkmark$|0.9450.9220.9140.8630.8440.9250.8550.8400.8090.7880.6500.832
(All labeled)|$\checkmark$|0.9570.9090.8790.9080.8640.9220.9010.8140.7140.8850.7020.817
|$\checkmark$||$\checkmark$|0.9150.8890.8850.7180.7110.8850.7800.7780.7680.5430.3630.749
|$\checkmark$|0.9110.8370.8550.7130.6560.8620.7730.6850.7120.5230.3850.712
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

For the validation dataset, which is part of the simulated dataset, a higher number of input dimensions were found to improve the results, enabling any classifier to classify the data with very high AUC. The best AUCs for all classified events are 0.993 and 0.995 for the ROC and precision-recall curve, respectively.

For the test dataset, the classification performance was verified using 1332 HSC SNe (1256 with redshift) labeled by the SALT2 light curve fitter (Guy et al. 2007, 2010), a conventional classification method. The verification label for the HSC SNe conforms to that reported in Yasuda et al. (2019), which defines SN Ia as SNe that satisfy all four of the following criteria for the SALT2 fitting results: (1) color (c) and stretch (x1) within the 3 σ range of the Scolnic and Kessler (2016) “All G10” distribution; (2) absolute magnitude in the B band, MB, brighter than −18.5 mag; (3) reduced χ2 of less than 10; (4) number of degrees of freedom (dof) greater than or equal to five. Other candidates that satisfy the looser set of conditions above were labeled “Ia?.” Specifically, the range in (1) was expanded to within 5 σ, and the thresholds of (2) and (3) were set to −17.5 mag and 20, respectively. Meanwhile, we defined non-SN Ia in the HSC classification as SNe that do not satisfy the conditions for classification as “Ia” or “Ia?,” and of which the number of dof is five or more. The number of labeled Ia, Ia?, and non-Ia are 429, 251, and 908 (410, 240, and 850, with redshift), respectively. Apart from the above, 215 SNe with less than five dof were labeled as “unclassified,” and the remaining 21 SNe failed to be fitted. This performance evaluation was conducted using 428 SNe Ia and 904 non-SNe Ia classified by our machine.

We also extracted 441 “light-curve-verified SNe” that have photometric information before and after their peak and for which spec-|$z$| are available, and verified their classification results. Figures 7 and 8 show the AUCs of the best classifier for all labeled HSC SNe and the light-curve-verified HSC SNe respectively. The confusion matrices for each case are shown in figure 9. The best-performing classifier obtained the same classification results as the conventional method for 84.2% of 1256 labeled SNe, which is 91.8% accurate for the 441 light-curve-verified SNe.

ROC and precision-recall curves for the two-class classification of all labeled HSC SNe. The input to the classifier is the pseudo-absolute magnitude and normalized flux. The colored lines represent the performance for each of the five classifiers with different input cases, and that for all of their outputs. (Color online)
Fig. 7.

ROC and precision-recall curves for the two-class classification of all labeled HSC SNe. The input to the classifier is the pseudo-absolute magnitude and normalized flux. The colored lines represent the performance for each of the five classifiers with different input cases, and that for all of their outputs. (Color online)

As figure 7, but for the light-curve-verified HSC SNe. (Color online)
Fig. 8.

As figure 7, but for the light-curve-verified HSC SNe. (Color online)

Normalized confusion matrices for the binary classification of 1256 labeled HSC SNe (left) and the 441 light-curve-verified SNe (right). The inputs for both classifications are the pseudo-absolute magnitude and normalized flux. (Color online)
Fig. 9.

Normalized confusion matrices for the binary classification of 1256 labeled HSC SNe (left) and the 441 light-curve-verified SNe (right). The inputs for both classifications are the pseudo-absolute magnitude and normalized flux. (Color online)

In the binary classification of 1256 labeled HSC SNe, 198 of them were misclassified. The misclassification rate for each case is different, and tends to increase as the number of input dimensions decreases, i.e., even though the rate is 13% for Case 0, it is 23% for Case 4. As with the PLAsTiCC data, incomplete events without their peak phase constitute the majority of misclassified events in the HSC data, accounting for 47% (93/198) of them. The second most common cause of misclassification is an outlier value or systematic flux offset in photometric data, accounting for 34% (67/198) of misclassifications. Of the remaining 38 SNe, 17 are boundary events with a Ia probability of 40% to 60%, and the remainder are events for which SALT2 fitting is ineffective.

 

5.2 Multi-type classification

In this paper we present the classification performance only for the validation dataset with the three-class classifier, because these three types of classification labels are not

 

 

 

available for the HSC transients. The accuracy values for each input in the three-class classification of the validation dataset are summarized in table 8. The best accuracy for the validation dataset is 94.0%. The confusion matrix of the best classifier is shown in figure 10. The result indicates that our classifier has a very high sensitivity toward SN Ia, whereas it is less effective at classifying SN Ibc.

Normalized confusion matrix for validation dataset in the HSC three-class classification. The input is the pseudo-absolute magnitude and normalized flux. The proportions in each row sum to 1 (within the rounding error). (Color online)
Fig. 10.

Normalized confusion matrix for validation dataset in the HSC three-class classification. The input is the pseudo-absolute magnitude and normalized flux. The proportions in each row sum to 1 (within the rounding error). (Color online)

Table 8.

Accuracy of each input in the HSC three-class classification for the validation dataset.

Input*Accuracy
MmfCase 01234All
|$\checkmark$||$\checkmark$|0.9850.9260.9200.8900.7740.940
|$\checkmark$|0.9710.8940.8860.8440.7290.914
|$\checkmark$||$\checkmark$|0.9700.9070.8970.8600.7240.921
|$\checkmark$|0.9520.8710.8610.8180.7010.892
Input*Accuracy
MmfCase 01234All
|$\checkmark$||$\checkmark$|0.9850.9260.9200.8900.7740.940
|$\checkmark$|0.9710.8940.8860.8440.7290.914
|$\checkmark$||$\checkmark$|0.9700.9070.8970.8600.7240.921
|$\checkmark$|0.9520.8710.8610.8180.7010.892
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

Accuracy of samples extracted from each case according to the fractions in table 4.

Table 8.

Accuracy of each input in the HSC three-class classification for the validation dataset.

Input*Accuracy
MmfCase 01234All
|$\checkmark$||$\checkmark$|0.9850.9260.9200.8900.7740.940
|$\checkmark$|0.9710.8940.8860.8440.7290.914
|$\checkmark$||$\checkmark$|0.9700.9070.8970.8600.7240.921
|$\checkmark$|0.9520.8710.8610.8180.7010.892
Input*Accuracy
MmfCase 01234All
|$\checkmark$||$\checkmark$|0.9850.9260.9200.8900.7740.940
|$\checkmark$|0.9710.8940.8860.8440.7290.914
|$\checkmark$||$\checkmark$|0.9700.9070.8970.8600.7240.921
|$\checkmark$|0.9520.8710.8610.8180.7010.892
*

Input to classifier is displayed as a check mark. M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

Accuracy of samples extracted from each case according to the fractions in table 4.

In addition, we describe the predicted classes of actual HSC SNe classified by the three-class classifier. Figure 11 shows the fractions of each type predicted by the classifier in each redshift from 0.1 to 1.5. All of the classified HSC SNe were used to calculate the fraction. These SN types are a combination of the outputs from the two classifiers with different inputs depending on the presence of redshift information: (1) pseudo-absolute magnitude and normalized flux; (2) magnitude and normalized flux.

Type fractions along redshift in HSC three-class classification. (Color online)
Fig. 11.

Type fractions along redshift in HSC three-class classification. (Color online)

5.3 Classification of HSC SNe

We report the classification results of 1824 HSC SNe, obtained by the proposed classifiers, in e-table 1.4 Part of this classification list is provided in table 9 as an example. This list summarizes the probabilities predicted by the two-class and three-class classifiers for each SN, along with the redshifts of the host galaxies and the classification labels assigned on the basis of the SALT2 fitting. The probabilities in this list are calculated from the output of the classifier with the normalized flux added to the input. Each classification performance shown in subsections 5.1 and 5.2 is calculated based on the probabilities in this list.

Table 9.

Example of classification result list for HSC SNe.

NameCase|$z$||$z$|_src* SALT2 fitting Classifier (Input: M + f) Classifier (Input: m + f)
dofTypeF_cover§2-class3-class2-class3-class
IaIaIbcIITypeIaIaIbcIIType
HSC16aaau1|$0.370_{-0.072}^{+0.110}$|37Ia?False0.5560.5540.0220.424Ia0.5170.6490.0080.342Ia
HSC16aaav1|$3.280_{-2.423}^{+0.167}$|417nonIaTrue0.1340.0490.0020.949II0.2790.3560.0480.596II
HSC16aabj0|$0.361_{-0.008}^{+0.007}$|28nonIaFalse0.6300.6670.0010.331Ia0.5740.5780.0180.405Ia
HSC16aabk109Ia?False0.4330.6750.0770.248Ia
HSC16aabp1|$1.477_{-0.032}^{+0.037}$|219nonIaFalse0.9570.9640.0010.035Ia0.8070.8710.0390.090Ia
HSC17bjrb1|$0.560_{-0.036}^{+0.024}$|31UCFalse0.0030.0070.0040.989II0.0110.0020.0040.994II
HSC17bjwo0|$1.449_{-0.063}^{+0.080}$|226IaTrue0.8810.9150.0050.080Ia0.8910.9350.0100.055Ia
HSC17bjya0|$1.128_{-0.000}^{+0.000}$|122nonIaTrue0.1300.1450.0390.816II0.1410.1090.0560.835II
HSC17bjyn0|$0.626_{-0.000}^{+0.000}$|124IaTrue0.8870.8910.0310.078Ia0.9650.9180.0070.075Ia
HSC17bjza1|$1.350_{-0.156}^{+1.142}$|413nonIaTrue0.0160.0410.0160.943II0.0620.0390.0050.957II
HSC17bkbn0|$0.863_{-0.012}^{+0.036}$|223nonIaTrue0.0310.0250.0020.973II0.0280.0210.0020.976II
HSC17bkcz0|$0.795_{-0.000}^{+0.000}$|127IaTrue0.6750.6740.0350.291Ia0.6610.7890.0190.191Ia
HSC17bkef0|$2.940_{-0.087}^{+0.119}$|20fail0.2190.4430.0000.556II0.9500.9470.0100.043Ia
HSC17bkem2|$0.609_{-0.000}^{+0.000}$|117IaTrue0.8890.8580.0010.141Ia0.9010.8630.0230.114Ia
HSC17bkfv0|$0.670_{-0.035}^{+0.035}$|323IaTrue0.9150.9060.0160.078Ia0.9610.9260.0110.063Ia
HSC17dskd0|$0.630_{-0.000}^{+0.000}$|13UCFalse0.8890.8630.0870.050Ia0.8730.8730.0720.054Ia
HSC17dsng0|$1.331_{-0.048}^{+0.048}$|27Ia?False0.9510.9670.0060.027Ia0.9350.8950.0110.094Ia
HSC17dsoh0|$1.026_{-0.000}^{+0.000}$|12UCFalse0.9680.9680.0110.020Ia0.9110.9230.0220.055Ia
HSC17dsox0|$1.137_{-0.034}^{+0.041}$|22UCFalse0.7080.7940.0190.186Ia0.7210.7380.0400.222Ia
HSC17dspl0|$0.624_{-0.000}^{+0.000}$|19nonIaFalse0.1800.0650.1140.821II0.0490.1030.1000.797II
NameCase|$z$||$z$|_src* SALT2 fitting Classifier (Input: M + f) Classifier (Input: m + f)
dofTypeF_cover§2-class3-class2-class3-class
IaIaIbcIITypeIaIaIbcIIType
HSC16aaau1|$0.370_{-0.072}^{+0.110}$|37Ia?False0.5560.5540.0220.424Ia0.5170.6490.0080.342Ia
HSC16aaav1|$3.280_{-2.423}^{+0.167}$|417nonIaTrue0.1340.0490.0020.949II0.2790.3560.0480.596II
HSC16aabj0|$0.361_{-0.008}^{+0.007}$|28nonIaFalse0.6300.6670.0010.331Ia0.5740.5780.0180.405Ia
HSC16aabk109Ia?False0.4330.6750.0770.248Ia
HSC16aabp1|$1.477_{-0.032}^{+0.037}$|219nonIaFalse0.9570.9640.0010.035Ia0.8070.8710.0390.090Ia
HSC17bjrb1|$0.560_{-0.036}^{+0.024}$|31UCFalse0.0030.0070.0040.989II0.0110.0020.0040.994II
HSC17bjwo0|$1.449_{-0.063}^{+0.080}$|226IaTrue0.8810.9150.0050.080Ia0.8910.9350.0100.055Ia
HSC17bjya0|$1.128_{-0.000}^{+0.000}$|122nonIaTrue0.1300.1450.0390.816II0.1410.1090.0560.835II
HSC17bjyn0|$0.626_{-0.000}^{+0.000}$|124IaTrue0.8870.8910.0310.078Ia0.9650.9180.0070.075Ia
HSC17bjza1|$1.350_{-0.156}^{+1.142}$|413nonIaTrue0.0160.0410.0160.943II0.0620.0390.0050.957II
HSC17bkbn0|$0.863_{-0.012}^{+0.036}$|223nonIaTrue0.0310.0250.0020.973II0.0280.0210.0020.976II
HSC17bkcz0|$0.795_{-0.000}^{+0.000}$|127IaTrue0.6750.6740.0350.291Ia0.6610.7890.0190.191Ia
HSC17bkef0|$2.940_{-0.087}^{+0.119}$|20fail0.2190.4430.0000.556II0.9500.9470.0100.043Ia
HSC17bkem2|$0.609_{-0.000}^{+0.000}$|117IaTrue0.8890.8580.0010.141Ia0.9010.8630.0230.114Ia
HSC17bkfv0|$0.670_{-0.035}^{+0.035}$|323IaTrue0.9150.9060.0160.078Ia0.9610.9260.0110.063Ia
HSC17dskd0|$0.630_{-0.000}^{+0.000}$|13UCFalse0.8890.8630.0870.050Ia0.8730.8730.0720.054Ia
HSC17dsng0|$1.331_{-0.048}^{+0.048}$|27Ia?False0.9510.9670.0060.027Ia0.9350.8950.0110.094Ia
HSC17dsoh0|$1.026_{-0.000}^{+0.000}$|12UCFalse0.9680.9680.0110.020Ia0.9110.9230.0220.055Ia
HSC17dsox0|$1.137_{-0.034}^{+0.041}$|22UCFalse0.7080.7940.0190.186Ia0.7210.7380.0400.222Ia
HSC17dspl0|$0.624_{-0.000}^{+0.000}$|19nonIaFalse0.1800.0650.1140.821II0.0490.1030.1000.797II
*

Code for redshift source. 1: spec-|$z$|⁠; 2: COSMOS photo-|$z$|⁠; 3: HSC photo-|$z$| Ultra-Deep; 4: HSC photo-|$z$| Deep; 0: hostless.

M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

SN type labeled by SALT2 fitting; UC: unclassified.

§

Flag indicating whether the photometric data cover the period of 10 d before and 20 d after the peak. SNe with this flag set to False are defined as “incomplete events.”

Table 9.

Example of classification result list for HSC SNe.

NameCase|$z$||$z$|_src* SALT2 fitting Classifier (Input: M + f) Classifier (Input: m + f)
dofTypeF_cover§2-class3-class2-class3-class
IaIaIbcIITypeIaIaIbcIIType
HSC16aaau1|$0.370_{-0.072}^{+0.110}$|37Ia?False0.5560.5540.0220.424Ia0.5170.6490.0080.342Ia
HSC16aaav1|$3.280_{-2.423}^{+0.167}$|417nonIaTrue0.1340.0490.0020.949II0.2790.3560.0480.596II
HSC16aabj0|$0.361_{-0.008}^{+0.007}$|28nonIaFalse0.6300.6670.0010.331Ia0.5740.5780.0180.405Ia
HSC16aabk109Ia?False0.4330.6750.0770.248Ia
HSC16aabp1|$1.477_{-0.032}^{+0.037}$|219nonIaFalse0.9570.9640.0010.035Ia0.8070.8710.0390.090Ia
HSC17bjrb1|$0.560_{-0.036}^{+0.024}$|31UCFalse0.0030.0070.0040.989II0.0110.0020.0040.994II
HSC17bjwo0|$1.449_{-0.063}^{+0.080}$|226IaTrue0.8810.9150.0050.080Ia0.8910.9350.0100.055Ia
HSC17bjya0|$1.128_{-0.000}^{+0.000}$|122nonIaTrue0.1300.1450.0390.816II0.1410.1090.0560.835II
HSC17bjyn0|$0.626_{-0.000}^{+0.000}$|124IaTrue0.8870.8910.0310.078Ia0.9650.9180.0070.075Ia
HSC17bjza1|$1.350_{-0.156}^{+1.142}$|413nonIaTrue0.0160.0410.0160.943II0.0620.0390.0050.957II
HSC17bkbn0|$0.863_{-0.012}^{+0.036}$|223nonIaTrue0.0310.0250.0020.973II0.0280.0210.0020.976II
HSC17bkcz0|$0.795_{-0.000}^{+0.000}$|127IaTrue0.6750.6740.0350.291Ia0.6610.7890.0190.191Ia
HSC17bkef0|$2.940_{-0.087}^{+0.119}$|20fail0.2190.4430.0000.556II0.9500.9470.0100.043Ia
HSC17bkem2|$0.609_{-0.000}^{+0.000}$|117IaTrue0.8890.8580.0010.141Ia0.9010.8630.0230.114Ia
HSC17bkfv0|$0.670_{-0.035}^{+0.035}$|323IaTrue0.9150.9060.0160.078Ia0.9610.9260.0110.063Ia
HSC17dskd0|$0.630_{-0.000}^{+0.000}$|13UCFalse0.8890.8630.0870.050Ia0.8730.8730.0720.054Ia
HSC17dsng0|$1.331_{-0.048}^{+0.048}$|27Ia?False0.9510.9670.0060.027Ia0.9350.8950.0110.094Ia
HSC17dsoh0|$1.026_{-0.000}^{+0.000}$|12UCFalse0.9680.9680.0110.020Ia0.9110.9230.0220.055Ia
HSC17dsox0|$1.137_{-0.034}^{+0.041}$|22UCFalse0.7080.7940.0190.186Ia0.7210.7380.0400.222Ia
HSC17dspl0|$0.624_{-0.000}^{+0.000}$|19nonIaFalse0.1800.0650.1140.821II0.0490.1030.1000.797II
NameCase|$z$||$z$|_src* SALT2 fitting Classifier (Input: M + f) Classifier (Input: m + f)
dofTypeF_cover§2-class3-class2-class3-class
IaIaIbcIITypeIaIaIbcIIType
HSC16aaau1|$0.370_{-0.072}^{+0.110}$|37Ia?False0.5560.5540.0220.424Ia0.5170.6490.0080.342Ia
HSC16aaav1|$3.280_{-2.423}^{+0.167}$|417nonIaTrue0.1340.0490.0020.949II0.2790.3560.0480.596II
HSC16aabj0|$0.361_{-0.008}^{+0.007}$|28nonIaFalse0.6300.6670.0010.331Ia0.5740.5780.0180.405Ia
HSC16aabk109Ia?False0.4330.6750.0770.248Ia
HSC16aabp1|$1.477_{-0.032}^{+0.037}$|219nonIaFalse0.9570.9640.0010.035Ia0.8070.8710.0390.090Ia
HSC17bjrb1|$0.560_{-0.036}^{+0.024}$|31UCFalse0.0030.0070.0040.989II0.0110.0020.0040.994II
HSC17bjwo0|$1.449_{-0.063}^{+0.080}$|226IaTrue0.8810.9150.0050.080Ia0.8910.9350.0100.055Ia
HSC17bjya0|$1.128_{-0.000}^{+0.000}$|122nonIaTrue0.1300.1450.0390.816II0.1410.1090.0560.835II
HSC17bjyn0|$0.626_{-0.000}^{+0.000}$|124IaTrue0.8870.8910.0310.078Ia0.9650.9180.0070.075Ia
HSC17bjza1|$1.350_{-0.156}^{+1.142}$|413nonIaTrue0.0160.0410.0160.943II0.0620.0390.0050.957II
HSC17bkbn0|$0.863_{-0.012}^{+0.036}$|223nonIaTrue0.0310.0250.0020.973II0.0280.0210.0020.976II
HSC17bkcz0|$0.795_{-0.000}^{+0.000}$|127IaTrue0.6750.6740.0350.291Ia0.6610.7890.0190.191Ia
HSC17bkef0|$2.940_{-0.087}^{+0.119}$|20fail0.2190.4430.0000.556II0.9500.9470.0100.043Ia
HSC17bkem2|$0.609_{-0.000}^{+0.000}$|117IaTrue0.8890.8580.0010.141Ia0.9010.8630.0230.114Ia
HSC17bkfv0|$0.670_{-0.035}^{+0.035}$|323IaTrue0.9150.9060.0160.078Ia0.9610.9260.0110.063Ia
HSC17dskd0|$0.630_{-0.000}^{+0.000}$|13UCFalse0.8890.8630.0870.050Ia0.8730.8730.0720.054Ia
HSC17dsng0|$1.331_{-0.048}^{+0.048}$|27Ia?False0.9510.9670.0060.027Ia0.9350.8950.0110.094Ia
HSC17dsoh0|$1.026_{-0.000}^{+0.000}$|12UCFalse0.9680.9680.0110.020Ia0.9110.9230.0220.055Ia
HSC17dsox0|$1.137_{-0.034}^{+0.041}$|22UCFalse0.7080.7940.0190.186Ia0.7210.7380.0400.222Ia
HSC17dspl0|$0.624_{-0.000}^{+0.000}$|19nonIaFalse0.1800.0650.1140.821II0.0490.1030.1000.797II
*

Code for redshift source. 1: spec-|$z$|⁠; 2: COSMOS photo-|$z$|⁠; 3: HSC photo-|$z$| Ultra-Deep; 4: HSC photo-|$z$| Deep; 0: hostless.

M: pseudo-absolute magnitude; m: magnitude; f: normalized flux.

SN type labeled by SALT2 fitting; UC: unclassified.

§

Flag indicating whether the photometric data cover the period of 10 d before and 20 d after the peak. SNe with this flag set to False are defined as “incomplete events.”

5.4 Dependence on the number of epochs

When using our classification method, the number of photometric data points given to the classifier increases as the survey progresses. Therefore, we investigated the transition of performance against the number of epochs. This was accomplished by classifying the HSC dataset by increasing the number of input data points in increments of one, and by examining the relationship between the number of epochs and the classification performance. Binary classifiers were adopted for classification, and the accuracy calculated from each confusion matrix was used for evaluation. Figure 12 shows the transition of classification performance for the Case 0 HSC dataset along with the number of epochs. Although the Ia accuracy is as low as 0.6 to 0.7 in the early stage of the survey with less than five epochs, it exceeds 0.8 when the number of epochs increases to 22. The partial decrease in accuracy is thought to be due to the new SNe being found upon the addition of a photometric point.

Relationship between the number of epochs and classification performance in binary classification for the Case 0 dataset. The horizontal axis represents the number of elapsed days of the HSC survey, and the vertical dotted line indicates the scale of the number of photometric points that were used as input. The color of each mark in accuracy indicates the band of the added photometric point. The horizontal dashed line indicates the accuracy when using all epochs. (Color online)
Fig. 12.

Relationship between the number of epochs and classification performance in binary classification for the Case 0 dataset. The horizontal axis represents the number of elapsed days of the HSC survey, and the vertical dotted line indicates the scale of the number of photometric points that were used as input. The color of each mark in accuracy indicates the band of the added photometric point. The horizontal dashed line indicates the accuracy when using all epochs. (Color online)

We also investigated the classification performance during each SN phase by regrouping all events according to the length of time since the first detection. Figure 13 illustrates the light curves and Ia probability transitions since the first detection of the three types of HSC SNe. We define “first detection” as the first day when the SN is detected with 5 σ confidence in flux, and which is flagged as a real object by the real–bogus classifier using a convolutional neural network (Yasuda et al. 2019). The probability is updated at each new epoch. Although the probability increases for certain events as the SN phase progresses, in the case of other events the probabilities fluctuate around 0.5 even as the observation progresses, and these events cannot be clearly classified. Figure 14 shows the relationship between the number of days since the first detection until classification, the accuracy, and the cumulative number of SNe classified as Ia with high probability. The calculations for each performance are based on the classification results for 1161 SNe that were detected before the rising phase. This figure presents the time span of SN photometric data that is needed for highly accurate classification using our classifier. The classification accuracy is 78.1% for the first two weeks of data, and after one month it increases to 82.7%. In addition, the number of follow-up candidates identified by the classifier can be estimated from the cumulative number in figure 14. Using data acquired within one month of the first detection, 79 SNe with |$z$| > 1 could be classified with Ia probability of 95% or more. Because the number of these SNe is cumulative, observed during a period of six months, dividing this by six corresponds to the number of follow-up SNe classified during a one-month survey, which is 13 events.

Examples of light curves and probability transitions. The title of each plot shows the name of the SN in the HSC survey and the label classified by SALT2 fitting. (Color online)
Fig. 13.

Examples of light curves and probability transitions. The title of each plot shows the name of the SN in the HSC survey and the label classified by SALT2 fitting. (Color online)

Classification accuracy (upper line) and cumulative number of Type Ia SNe classified with high probability against the SN phase. The middle line indicates the cumulative number of SNe with Ia probability >95%, and the lower line is that for distant SNe at $z$ > 1. (Color online)
Fig. 14.

Classification accuracy (upper line) and cumulative number of Type Ia SNe classified with high probability against the SN phase. The middle line indicates the cumulative number of SNe with Ia probability >95%, and the lower line is that for distant SNe at |$z$| > 1. (Color online)

Lastly, we studied the transition of the SN Ia probability along the SN phase. Each time the number of input epochs increases, the SN Ia probability, i.e., the output of the classifier, is updated. In the upper panel of figure 15 we plot the SN Ia probability at each epoch for each of 300 SNe labeled as Ia and non-Ia based on the SALT2 fitting. The use of this result enabled us to measure the last epoch at which the correct type is classified, i.e., the epoch after which no further change in the classification occurs. The lower panel shows the cumulative ratio for the epoch. The figure shows that the classification performance improves with time and that 80% of SNe are correctly classified approximately 30 d after their first detection. In this figure, the initial cumulative ratio is lower than the accuracy shown in figure 14 because certain SNe that are initially correctly classified could ultimately be misclassified as a wrong type.

Upper panel: Transition of Ia probability of SNe after first detection. Each line corresponds to one of the SNe. Different colors indicate different labels, with red being Ia and blue being non-Ia. Lower panel: Cumulative ratio for the epoch at which SNe are finally correctly classified. (Color online)
Fig. 15.

Upper panel: Transition of Ia probability of SNe after first detection. Each line corresponds to one of the SNe. Different colors indicate different labels, with red being Ia and blue being non-Ia. Lower panel: Cumulative ratio for the epoch at which SNe are finally correctly classified. (Color online)

Figure 16 shows the transitions of SN Ia probability as a function of the number of days from the peak for 26 SNe selected as HST targets in the HSC survey, and the average of these transitions. The SN Ia probability for the average of the candidates is greater than 0.8 three weeks before the peak. This means that our classifier accomplishes the task described in subsection 2.1 by identifying good candidate SNe even when it only has information acquired before the SN peak.

Ia probability transitions of 26 HST targets. Each of the lighter lines represents the variation for individual HST targets, and the thick line is the average of these lines. (Color online)
Fig. 16.

Ia probability transitions of 26 HST targets. Each of the lighter lines represents the variation for individual HST targets, and the thick line is the average of these lines. (Color online)

6 Discussion

6.1 Factors affecting classifier performance

In this study we applied our classifier to actual HSC survey data to evaluate its classification performance. By classifying actual data with our method, we determined that the performance difference between the validation and test sets is larger than that for the PLAsTiCC data. Here we discuss the factors that affect the classification performance of actual data.

One possible factor is the uncertainty in the labeling of actual data using conventional methods. In the confusion matrix for all labeled SNe in figure 9, 47% of the 198 misclassified SNe are incomplete events, and these events only form 35% of the 1824 HSC SNe. The high percentage of incomplete events among the misclassified events suggests that incomplete events reduce the classification performance. However, the misclassification rate of incomplete events is clearly higher at 28% (93/338) for the HSC data compared to 7% (58/871) for PLAsTiCC data. In addition, SALT2 fitting is equally ineffective at fitting incomplete events owing to its specification, and errors in the fitting parameters are significantly larger. These findings suggest a degree of uncertainty in labeling with conventional classification methods. Another source of uncertainty in labeling is the uncertainty in the redshift information used in SALT2 fitting. For at least 16 misclassified events that were not affected by any other misclassification factors, the fitting color parameters are far outside the criteria for Ia, and the wrong redshift was probably used for fitting. In fact, as shown in the panel on the right in figure 9, the performance for light-curve-verified SNe, except for incomplete events and for events for which spec-|$z$| is not known, is superior to that of all classified events.

Another reason for the large difference in the performance of the validation and test set is the inability to perfectly simulate the observed data. Outlier values and systematic flux offsets, which are considered to be one of the causes of misclassification as described in subsection 5.1, are found only in real observed data. As described in Yasuda et al. (2019), the photometric data of the SNe in the HSC survey are measured from the difference image obtained by subtracting the reference image from the observed image. We believe that the unsimulated residual in this subtraction created a difference between the photometric data that were used for training and observation, and increases the misclassification rate of the classifier when processing actual data. For example, the light curve of SN HSC16akvr shown in the panel on the right in figure 13 fluctuates at its tail, and in this part of the curve the classification is not clear. Improving the performance of the classifier would therefore necessitate the reduction or simulation of these outlier values as much as possible.

6.2 Comparison with other classifiers

The direct comparison of the classification results of the classifiers is complicated by differences in methods, training datasets, or the number of inputs and outputs. Here, we simply compare the results obtained with the recent SN type classifier based on machine learning and our classifier with AUC of ROC in binary classification to determine whether an SN is of type Ia. For comparison, we use the classification results of the simulated SN light curves with redshift information from Lochner et al. (2016), Charnock and Moss (2017), and Muthukrishna et al. (2019a). Lochner et al. (2016) obtained an AUC of 0.984 by using boosted decision trees for classification using the SALT2 fitting parameters as input features. Charnock and Moss (2017) reported an AUC of 0.986 for the classification of SNPCC data using deep recurrent neural networks (RNN) with unidirectional long short-term memory (LSTM) units. Muthukrishna et al. (2019a) used deep RNN with gated recurrent units (GRUs) to classify simulated ZTF light curves and achieved an AUC of 0.99 40 d after the trigger. The AUC of our classifier, 0.996, is comparable to those of these recent classifiers in binary classification.

Next, we compare the results of our classifiers with those that ranked first in the PLAsTiCC Kaggle competition (Malz et al. 2019). The best classifier in the competition (Boone 2019) was based on Light-GBM and trained with features extracted from photometric data modeled by Gaussian process regression. The dataset that was used to train this classifier included a total of 591410 light curves that were obtained by augmenting 100 new light curves under different observation conditions and different redshifts for each light curve of the original PLAsTiCC training set. The classifier classifies events into 15 classes, including variable objects other than SN such as microlensing events and active galactic nuclei. We used 2297 PLAsTiCC SN predictions, classified by both classifiers, for comparison. Because the number of output classes is different between our classifier and the best classifier, we divided the classification results into SN Ia and other, and performed the comparison as a binary classification. Figure 17 shows the confusion matrix of each classifier. Although a strict comparison between our classifier and the best PLAsTiCC classifier is not possible because of the different training datasets and number of output classes, the capability of our classifier is comparable to that of the best PLAsTiCC classifier. However, because our method fixes the observation schedule to the type of input, it is impossible to process all PLAsTiCC data with one classifier. Our method is not necessarily suitable for surveys that sweep a wide area such as the LSST survey; instead, it is useful for surveys that observe the same field for a period of time, such as the HSC-SSP Transient Survey and LSST DDF.

Normalized confusion matrices of two-class classification results for (left) the PLAsTiCC test set by Boone (2019) and (right) our best classifier given the pseudo-absolute magnitude and normalized flux. (Color online)
Fig. 17.

Normalized confusion matrices of two-class classification results for (left) the PLAsTiCC test set by Boone (2019) and (right) our best classifier given the pseudo-absolute magnitude and normalized flux. (Color online)

6.3 Data uncertainty

Information about the flux error is also important for the classification of distant SN Ia, which is the main target of the HSC survey and has a small signal-to-noise ratio. As mentioned in subsection 4.3, the signal-to-noise ratio affects the performance of the model. Here, we discuss the incorporation of errors for classification, including in our method. Charnock and Moss (2017) added the flux error itself as part of the input vector. In the case of feature-based classification, the flux error is used in the fitting process to calculate the features, and affects those errors. The flux error is also used to calculate statistical features such as “standard deviation/mean,” which is one of the features discussed by Narayan et al. (2018) and Muthukrishna et al. (2019a). As described in subsection 3.2, we calculate the flux error of the simulated training data based on the relationship between the flux and the flux error for each epoch of the observed data. Then, random numbers that follow a normal distribution defined by the original fluxes and flux errors are input to the classifier as processed fluxes.

The processing of time series data with uncertainty is problematic in many application domains, not only in the field of astronomy. For example, in the similarity matching of time series data with uncertainty, considering continuous correlation in time series such as measuring the distance after filtering by the uncertain moving average (UMA; Dallachiesa et al. 2012) is suggested as a promising approach. UMA filtering uses the correlation of neighboring points and reduces the contribution of observation points with large errors. We consider the development of a method to efficiently incorporate this error information to be necessary for improving the classification performance.

6.4 Missing data

Our method was developed specifically for the HSC survey, where each observed transient tends to have the same cadence by observing a certain field repeatedly. However, this method is neither suitable for processing data with different observation schedules for each object, nor for datasets from which data are missing. In other classification methods, missing data are interpolated by linear or Gaussian processes (Lochner et al. 2016; Muthukrishna et al. 2019a), or replaced with reference to the flux of adjacent photometric points (Charnock & Moss 2017). Our approach is to handle missing data by preparing multiple classifiers. The advantage of our classifier is that it is able to use the photometric information directly for classification with less preprocessing, compared to methods that use interpolation or replacement. In the case of the HSC survey, most of the observation schedules of all SNe can be roughly divided into two types that correspond to either the Ultra-Deep or Deep layers. Therefore, we can classify 99.3% of SNe using only five classifiers with different input schedules, including those for SNe data containing missing data (table 4).

6.5 Type fractions of HSC SNe

The elemental abundance of the solar system (Grevesse & Sauval 1998) originates from the cosmic history of SNe (Kobayashi et al. 2000; Maraston 2005). Recent studies have shown that the solar abundance pattern is observed in other systems (Ramírez et al. 2009) and clusters (Mernier et al. 2018). Investigation of the origin of elements in the context of cosmic evolution is thus important (Fukugita & Peebles 2004).

It is now well established that the star formation rate peaks at |$z$| ∼ 2, and that the chemical composition of our system is a mixture of SN Ia and core-collapse SNe (Tsujimoto et al. 1995; Kobayashi & Nakasato 2011). Deriving the SN rate would require careful analysis (Dilday et al. 2008; Brown et al. 2019; Frohmaier et al. 2019) and is beyond the scope of this paper, but at least we can verify the consistency with previous work in terms of the relative fraction. The Lick Observatory Supernova Survey (Li et al. 2011) reported the relative ratio of SN Ia : Ibc : II to be 0.24 : 0.19 : 0.57, whereas we obtained 0.22 : 0.19 : 0.59 even at |$z$| ≲ 0.2.

Based on our survey depth, we have a complete sampling of SN Ia up to |$z$| ∼ 1.1, although we lose the completeness of core-collapse SN in much lower redshift given the fact that the magnitudes of SN Ibc and SN II are fainter by 2–3 mag at maximum. Increasing redshift causes the SN II fraction to decrease whereas the SN Ia fraction increases, as shown in figure 11. This completeness effect is simply due to the magnitude difference and does not reflect the cosmological SN rates. By adopting the SN Ia rate from Graur et al. (2014) and the core-collapse SN rate from Strolger et al. (2015), we can estimate the completeness of core-collapse SN. At |$z$| ∼ 0.3, the core-collapse SN completeness is 78%, and it is reduced to 49% at |$z$| ∼ 0.5. The reason for the SN II fraction not approaching zero at |$z$| ∼ 1 in figure 11 is that the magnitude of the dispersion of core-collapse SN (Li et al. 2011; Kessler et al. 2019; σ ∼ 1.2 mag) is much larger than that of SN Ia (Rubin et al. 2015; σ ∼ 0.5 mag), and they are more abundant at |$z$| ∼ 1 by a factor of 4–5 (Madau et al. 1998; Hounsell et al. 2018). Although a more careful investigation is necessary, we deduce the completeness of core-collapse SN at |$z$| ∼ 1 to be 12%.

7 Conclusions

We have presented a model of a classifier that classifies SN types by directly accepting photometric data as input. The classification performance of the classifier was discussed in detail. Our DNN classifier is trained with simulated SN photometric data and was shown to classify PLAsTiCC data and actual HSC SNe data with high accuracy of 95.3% and 84.2%, respectively. Our study of the number of input dimensions also indicated that our classifier can classify the HSC survey data with sufficient accuracy even using two weeks of pre-peak data since the first detection. Based on these results, we conclude that this classifier has sufficient classification performance for subsequent type-specific studies and for the selection of follow-up targets even in actual HSC surveys.

The following supplementary data is available at PASJ online.

E-table 1

Acknowledgements

The Hyper Suprime-Cam (HSC) collaboration includes the astronomical communities of Japan and Taiwan, and Princeton University. The HSC instrumentation and software were developed by the National Astronomical Observatory of Japan (NAOJ), the Kavli Institute for the Physics and Mathematics of the Universe (Kavli IPMU), the University of Tokyo, the High Energy Accelerator Research Organization (KEK), the Academia Sinica Institute for Astronomy and Astrophysics in Taiwan (ASIAA), and Princeton University. Funding was contributed by the FIRST program from the Japanese Cabinet Office, the Ministry of Education, Culture, Sports, Science and Technology (MEXT), the Japan Society for the Promotion of Science (JSPS), the Japan Science and Technology Agency (JST), the Toray Science Foundation, NAOJ, Kavli IPMU, KEK, ASIAA, and Princeton University.

This paper makes use of software developed for the Large Synoptic Survey Telescope. We thank the LSST Project for making their code available as free software at 〈http://dm.lsst.org〉.

The Pan-STARRS1 Surveys (PS1) have been made possible through contributions of the Institute for Astronomy, the University of Hawaii, the Pan-STARRS Project Office, the Max-Planck Society and its participating institutes, the Max Planck Institute for Astronomy, Heidelberg and the Max Planck Institute for Extraterrestrial Physics, Garching, The Johns Hopkins University, Durham University, the University of Edinburgh, Queen’s University Belfast, the Harvard-Smithsonian Center for Astrophysics, the Las Cumbres Observatory Global Telescope Network Incorporated, the National Central University of Taiwan, the Space Telescope Science Institute, the National Aeronautics and Space Administration under Grant No. NNX08AR22G issued through the Planetary Science Division of the NASA Science Mission Directorate, the National Science Foundation under Grant No. AST-1238877, the University of Maryland, and Eotvos Lorand University (ELTE) and the Los Alamos National Laboratory.

We thank Y. Imoto for his enormous contribution to the development of the classification model, and the data classification. We also thank T. J. Moriya, J. Jiang, and other members of the HSC transient working group for helpful discussions and comments on the manuscript.

This work is supported by JST CREST Grant Number JPMHCR1414, JST AIP Acceleration Research Grant Number JP20317829, MEXT KAKENHI Grant Numbers 18H04345 (N.Ya.), 17H06363 (M.T.), and JSPS KAKENHI Grant Numbers 18K03696, 20H04730, 20HT0063 (N.S.), 16H02183 (M.T.), 19H00694 (M.T.), 20H00179 (N.T.).

This research is based in part on data collected at the Subaru Telescope and retrieved from the HSC data archive system, which is operated by the Subaru Telescope and Astronomy Data Center at NAOJ.

Appendix. Redshift estimation

The HSC-SSP Transient Survey includes SNe without redshift because their host galaxies are not clearly identified. These SNe are referred to as hostless SNe. In the redshift list of HSC SNe that was used for type classification, 6% (108/1824) correspond to these hostless SNe. To be able to process these SNe with our classifier and to address the possibility of host galaxy misidentification, we could estimate the redshift |$z$| of a SN from the photometric data.

The redshift was estimated using a model with the same structure as figure 3 except that the DNN output is a scalar, and the model does not include a final softmax layer for redshift estimation. The objective function to optimize is the squared error between the ground truth |$z$| and the output value |$\hat{y}$|⁠. We measured the accuracy of the model with the coefficient of determination R2; that is,
(A1)
where |$z$|n is the redshift value of the nth sample, |$\hat{y}_n$| is the output of the nth sample, and |$\bar{z}$| is the mean redshift of the dataset. We performed the same hyperparameter search in the model as in subsection 4.2, including data augmentation. The accuracy of this estimator is discussed below.

We applied the standard deviation of the normalized residual (⁠|$z$|pred|$z$|spec)/(1 + |$z$|spec) (Salvato et al. 2009, 2019), used in galaxy photo-|$z$| estimation, to the performance criteria for redshift estimation. Verified with Case 0 simulated data, the redshifts for Ia and non-Ia objects were estimated to be 0.022 and 0.076, respectively. For the actual HSC SNe, we used the observationally obtained redshifts of host galaxies (see section 5) including spec-|$z$| and photo-|$z$| for comparison with the estimates. Figure 18 shows a comparison between the estimated redshifts and the observed redshifts for HSC SNe classified as Ia. The two classes labeled by SALT2 fitting have different distributions, and the events labeled non-Ia are less accurate in terms of their redshift estimation than Ia. The normalized residuals for each of the Ia and non-Ia Case 0 SNe labeled with SALT2 fitting are normally distributed with standard deviations of 0.066 and 0.138, respectively. Limiting the comparison to SNe in the host galaxies with spec-|$z$| information, the standard deviations are 0.056 and 0.130, respectively. Although this estimation accuracy is lower than the template fitting accuracy using host galaxy photometric data, it is useful not only for hostless SNe, but also to select the host galaxy by comparing the redshift estimated from the SN light curve itself with those of the galaxies. Furthermore, this distributional difference of residuals leads to the fact that the Ia accuracy can be further improved by excluding events with large residuals when estimating the redshift of the SN and host galaxy.

(Left) Relation between the machine predictions and the observed redshifts for HSC SNe. The dashed line is the line of equality. The color difference represents the SALT2 fitted label. (Right) Normalized residual distribution for Case 0 SNe. (Color online)
Fig. 18.

(Left) Relation between the machine predictions and the observed redshifts for HSC SNe. The dashed line is the line of equality. The color difference represents the SALT2 fitted label. (Right) Normalized residual distribution for Case 0 SNe. (Color online)

Footnotes

2

The code for our model is available at 〈https://github.com/ichiro-takahashi/snclass〉.

4

E-table 1 is available in the online edition as a supplementary table.

References

Abadi
 
M.
 et al.  
2016
,

Aihara
 
H.
 et al.  
2018a
,
PASJ
,
70
,
S4

Aihara
 
H.
 et al.  
2018b
,
PASJ
,
70
,
S8

Appenzeller
 
I.
 et al.  
1998
,
Messenger
,
94
,
1

Bergstra
 
J.
,
Yamins
D.
,
Cox
D.
 
2013
, in
Proc. 30th Int. Conf. Machine Learning
, ed.
Dasgupta
S.
,
McAllester
D.
(
Cambridge, MA
:
PMLR
),
115

Betoule
 
M.
 et al.  
2014
,
A&A
,
568
,
A22

Bjorck
 
N.
,
Gomes
C. P.
,
Selman
B.
,
Weinberger
K. Q.
 
2018
, in
Proc. Neural Information Processing Systems 2018
, ed.
Bengio
S.
et al.
(
San Diego
:
Neural Information Processing Systems
),
7694

Boone
 
K.
 
2019
,
AJ
,
158
,
257

Brout
 
D.
 et al.  
2019
,
ApJ
,
874
,
150

Brown
 
J. S.
 et al.  
2019
,
MNRAS
,
484
,
3785

Cao
 
Y.
 et al.  
2015
,
Nature
,
521
,
328

Carliles
 
S.
,
Budavári
T.
,
Heinis
S.
,
Priebe
C.
,
Szalay
A. S.
 
2010
,
ApJ
,
712
,
511

Carrasco-Davis
 
R.
 et al.  
2020
,

Charnock
 
T.
,
Moss
A.
 
2017
,
ApJ
,
837
,
L28

Coil
 
A. L.
 et al.  
2011
,
ApJ
,
741
,
8

Conley
 
A.
 et al.  
2006
,
AJ
,
132
,
1707

Collister
 
A. A.
,
Lahav
O.
 
2004
,
PASP
,
116
,
345

Cooke
 
J.
 et al.  
2012
,
Nature
,
491
,
228

Curtin
 
C.
 et al.  
2019
,
ApJS
,
241
,
17

Dallachiesa
 
M.
,
Nushi
B.
,
Mirylenka
K.
,
Palpanas
T.
 
2012
,
Proc. VLDB Endow.
,
5
,
1662

de Jaeger
 
T.
 et al.  
2017
,
MNRAS
,
472
,
4233

Dilday
 
B.
 et al.  
2008
,
ApJ
,
682
,
262

Filippenko
 
A. V.
 
1997
,
ARA&A
,
35
,
309

Förster
 
F.
 et al.  
2018
,
Nature Astron.
,
2
,
808

Förster
 
F.
 et al.  
2020
,

Frohmaier
 
C.
 et al.  
2019
,
MNRAS
,
486
,
2308

Fukugita
 
M.
,
Peebles
P. J. E.
 
2004
,
ApJ
,
616
,
643

Furusawa
 
H.
 et al.  
2018
,
PASJ
,
70
,
S3

Gal-Yam
 
A.
 
2012
,
Science
,
337
,
927

Garcia-Dias
 
R.
,
Allende Prieto
C.
,
Sánchez Almeida
J.
,
Ordovás-Pascual
I.
 
2018
,
A&A
,
612
,
A98

Goldstein
 
D. A.
 et al.  
2015
,
AJ
,
150
,
82

Graur
 
O.
 et al.  
2014
,
ApJ
,
783
,
28

Grevesse
 
N.
,
Sauval
A. J.
 
1998
,
Space Sci. Rev.
,
85
,
161

Guy
 
J.
 et al.  
2007
,
A&A
,
466
,
11

Guy
 
J.
 et al.  
2010
,
A&A
,
523
,
A7

Hasinger
 
G.
 et al.  
2018
,
ApJ
,
858
,
77

Hausen
 
R.
,
Robertson
B. E.
 
2020
,
ApJS
,
248
,
20

Hook
 
I. M.
,
Jørgensen
I.
,
Allington-Smith
J. R.
,
Davies
R. L.
,
Metcalfe
N.
,
Murowinski
R. G.
,
Crampton
D.
 
2004
,
PASP
,
116
,
425

Hosseinzadeh
 
G.
 et al.  
2020
,

Hounsell
 
R.
 et al.  
2018
,
ApJ
,
867
,
23

Howell
 
D. A.
 et al.  
2006
,
Nature
,
443
,
308

Ioffe
 
S.
,
Szegedy
C.
 
2015
, in
Proc. 32nd Int. Conf. Machine Learning
, ed.
Bach
F.
,
Blei
D.
(
Cambridge, MA
:
PMLR
),
448

Ivezić
 
Ž.
 et al.  
2019
,
ApJ
,
873
,
111

Jiang
 
J.-A.
 et al.  
2017
,
Nature
,
550
,
80

Jiang
 
J.-A.
 et al.  
2020
,
ApJ
,
892
,
25

Jones
 
D. O.
 et al.  
2018
,
ApJ
,
857
,
51

Kashikawa
 
N.
 et al.  
2002
,
PASJ
,
54
,
819

Kawanomoto
 
S.
 et al.  
2018
,
PASJ
,
70
,
66

Kessler
 
R.
 et al.  
2019
,
PASP
,
131
,
094501

Kimura
 
A.
,
Takahashi
I.
,
Tanaka
M.
,
Yasuda
N.
,
Ueda
N.
,
Yoshida
N.
 
2017
, in
37th IEEE Distributed Computing Systems Workshops
, ed.
Musaev
A.
et al.
(
New York
:
IEEE Computer Society
),
354

Kingma
 
D. P.
,
Ba
J.
 
2014
,
arXiv:1412.6980

Kobayashi
 
C.
,
Nakasato
N.
 
2011
,
ApJ
,
729
,
16

Kobayashi
 
C.
,
Tsujimoto
T.
,
Nomoto
K.
 
2000
,
ApJ
,
539
,
26

Komiyama
 
Y.
 et al.  
2018
,
PASJ
,
70
,
S2

Laigle
 
C.
 et al.  
2016
,
ApJS
,
224
,
24

Li
 
W.
 et al.  
2011
,
MNRAS
,
412
,
1441

Li
 
X.
,
Chen
S.
,
Hu
X.
,
Yang
J.
 
2019
, in
2019 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR)
(
New York
:
IEEE Computer Society
),
2677

Linder
 
E. V.
 
2003
,
Phys. Rev. Lett.
,
90
,
091301

Lochner
 
M.
,
McEwen
J. D.
,
Peiris
H. V.
,
Lahav
O.
,
Winter
M. K.
 
2016
,
ApJS
,
225
,
31

Lupton
 
R. H.
,
Gunn
J. E.
,
Szalay
A. S.
 
1999
,
AJ
,
118
,
1406

Madau
 
P.
,
della Valle
M.
,
Panagia
N.
 
1998
,
MNRAS
,
297
,
L17

Maeda
 
K.
,
Jiang
J.-a.
,
Shigeyama
T.
,
Doi
M.
 
2018
,
ApJ
,
861
,
78

Malz
 
A. I.
 et al.  
2019
,
AJ
,
158
,
171

Maraston
 
C.
 
2005
,
MNRAS
,
362
,
799

Masters
 
D. C.
,
Stern
D. K.
,
Cohen
J. G.
,
Capak
P. L.
,
Rhodes
J. D.
,
Castander
F. J.
,
Paltani
S.
 
2017
,
ApJ
,
841
,
111

Mernier
 
F.
 et al.  
2018
,
Space Sci. Rev.
,
214
,
129

Miyazaki
 
S.
 et al.  
2018
,
PASJ
,
70
,
S1

Morii
 
M.
 et al.  
2016
,
PASJ
,
68
,
104

Moriya
 
T. J.
 et al.  
2019
,
ApJS
,
241
,
16

Moriya
 
T. J.
,
Sorokina
E. I.
,
Chevalier
R. A.
 
2018
,
Space Sci. Rev.
,
214
,
59

Mosher
 
J.
 et al.  
2014
,
ApJ
,
793
,
16

Muthukrishna
 
D.
,
Narayan
G.
,
Mandel
K. S.
,
Biswas
R.
,
Hložek
R.
 
2019a
,
PASP
,
131
,
118002

Muthukrishna
 
D.
,
Parkinson
D.
,
Tucker
B. E.
 
2019b
,
ApJ
,
885
,
85

Narayan
 
G.
 et al.  
2018
,
ApJS
,
236
,
9

Oke
 
J. B.
 et al.  
1995
,
PASP
,
107
,
375

Papadogiannakis
 
S.
 et al.  
2019
,
MNRAS
,
483
,
5045

Pasquet
 
J.
,
Bertin
E.
,
Treyer
M.
,
Arnouts
S.
,
Fouchez
D.
 
2019
,
A&A
,
621
,
A26

Perlmutter
 
S.
 et al.  
1999
,
ApJ
,
517
,
565

Petrillo
 
C. E.
 et al.  
2017
,
MNRAS
,
472
,
1129

Phillips
 
M. M.
 et al.  
2007
,
PASP
,
119
,
360

Quimby
 
R. M.
 et al.  
2011
,
Nature
,
474
,
487

Quimby
 
R. M.
,
Aldering
G.
,
Wheeler
J. C.
,
Höflich
P.
,
Akerlof
C. W.
,
Rykoff
E. S.
 
2007
,
ApJ
,
668
,
L99

Ramírez
 
I.
,
Meléndez
J.
,
Asplund
M.
 
2009
,
A&A
,
508
,
L17

Riess
 
A. G.
 et al.  
1998
,
AJ
,
116
,
1009

Rubin
 
D.
 et al.  
2015
,
ApJ
,
813
,
137

Sako
 
M.
 et al.  
2011
,
ApJ
,
738
,
162

Salvato
 
M.
 et al.  
2009
,
ApJ
,
690
,
1250

Salvato
 
M.
,
Ilbert
O.
,
Hoyle
B.
 
2019
,
Nature Astron.
,
3
,
212

Sánchez-áez
 
P.
 et al.  
2020
,

Saunders
 
W.
 et al.  
2004
, in
Proc. SPIE, 5492, Ground-based Instrumentation for Astronomy
, ed.
Moorwood
A. F. M.
&
Masanori
M.
(
Bellingham, WA
:
SPIE
),
389

Scolnic
 
D.
,
Kessler
R.
 
2016
,
ApJ
,
822
,
L35

Scolnic
 
D.
 et al.  
2014
,
ApJ
,
795
,
45

Scolnic
 
D. M.
 et al.  
2018
,
ApJ
,
859
,
101

Scoville
 
N.
 et al.  
2007
,
ApJS
,
172
,
1

Sharma
 
K.
,
Kembhavi
A.
,
Kembhavi
A.
,
Sivarani
T.
,
Abraham
S.
,
Vaghmare
K.
 
2020
,
MNRAS
,
491
,
2280

Silverman
 
J. D.
 et al.  
2015
,
ApJS
,
220
,
12

Srivastava
 
N.
,
Hinton
G.
,
Krizhevsky
A.
,
Sutskever
I.
,
Salakhutdinov
R.
 
2014
,
J. Mach. Learn. Res.
,
15
,
1929

Srivastava
 
R. K.
,
Greff
K.
,
Schmidhuber
J.
 
2015a
,

Srivastava
 
R. K.
,
Greff
K.
,
Schmidhuber
J.
 
2015b
,

Strolger
 
L.-G.
 et al.  
2015
,
ApJ
,
813
,
93

Tampo
 
Y.
 et al.  
2020
,
ApJ
,
894
,
27

Tanaka
 
M.
 et al.  
2018
,
PASJ
,
70
,
S9

The PLAsTiCC team
 
2018
,

Thompson
 
T. A.
,
Burrows
A.
,
Pinto
P. A.
 
2003
,
ApJ
,
592
,
434

Tominaga
 
N.
,
Morokuma
T.
,
Blinnikov
S. I.
,
Baklanov
P.
,
Sorokina
E. I.
,
Nomoto
K.
 
2011
,
ApJS
,
193
,
20

Tsujimoto
 
T.
,
Nomoto
K.
,
Yoshii
Y.
,
Hashimoto
M.
,
Yanagida
S.
,
Thielemann
F. K.
 
1995
,
MNRAS
,
277
,
945

Villar
 
V. A.
 et al.  
2020
,

Yasuda
 
N.
 et al.  
2019
,
PASJ
,
71
,
74

Zhang
 
H.
,
Cissé
M.
,
Dauphin
Y. N.
,
Lopez-Paz
D.
 
2017
,

This article is published and distributed under the terms of the Oxford University Press, Standard Journals Publication Model (https://dbpia.nl.go.kr/journals/pages/open_access/funder_policies/chorus/standard_publication_model)

Supplementary data