(Translated by https://www.hiragana.jp/)
Machine-learning Application to Fermi-LAT Data: Sharpening All-sky Map and Emphasizing Variable Sources - IOPscience

A publishing partnership

Machine-learning Application to Fermi-LAT Data: Sharpening All-sky Map and Emphasizing Variable Sources

, , , , , , , , and

Published 2021 May 27 © 2021. The American Astronomical Society. All rights reserved.
, , Citation Shogo Sato et al 2021 ApJ 913 83 DOI 10.3847/1538-4357/abf48f

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

0004-637X/913/2/83

Abstract

A novel application of machine-learning (ML) based image processing algorithms is proposed to analyze an all-sky map (ASM) obtained using the Fermi Gamma-ray Space Telescope. An attempt was made to simulate a 1 yr ASM from a short-exposure ASM generated from 1-week observation by applying three ML-based image processing algorithms: dictionary learning, U-net, and Noise2Noise. Although the inference based on ML is less clear compared to standard likelihood analysis, the quality of the ASM was generally improved. In particular, the complicated diffuse emission associated with the galactic plane was successfully reproduced only from 1-week observation data to mimic a ground truth (GT) generated from a 1 yr observation. Such ML algorithms can be implemented relatively easily to provide sharper images without various assumptions of emission models. In contrast, large deviations between simulated ML maps and the GT map were found, which are attributed to the significant temporal variability of blazar-type active galactic nuclei (AGNs) over a year. Thus, the proposed ML methods are viable not only to improve the image quality of an ASM but also to detect variable sources, such as AGNs, algorithmically, i.e., without human bias. Moreover, we argue that this approach is widely applicable to ASMs obtained by various other missions; thus, it has the potential to examine giant structures and transient events, both of which are rarely found in pointing observations.

Export citation and abstract BibTeX RIS

1. Introduction

All-sky monitoring in astronomy has begun a new era in which continuous monitoring of persistent sources, as well as automated searches to discover new transients, is possible. Moreover, the all-sky monitor is a unique approach that reveals the giant structures that are rarely seen in pointing observations. In the 1990s, ROSAT provided the first precise all-X-ray sky map below 2 keV (Snowden et al. 1997). The observation started in 1990 July and lasted for approximately 6 months. The RXTE-ASM (Bradt et al. 1993) and MAXI on board the International Space Station (Matsuoka et al. 2009) carry one-dimensional slat/shadow cameras extending the maximum energy window to 10 keV, which also makes it possible to monitor the brightest X-ray sources on a daily or hourly basis. More recently, eROSITA on the Spectrum Roentgen Gamma (Merloni et al. 2012) observatory, launched in 2019 June, has provided a new, sharp view of hot and energetic processes taking place across the universe. For gamma rays, the Fermi Gamma-ray Space Telescope was launched in 2008 (Atwood et al. 2009). It can observe the 30 MeV to 100 GeV gamma-ray sky with unprecedented sensitivity. Owing to a large field of view of 2.4 sr, the large-area telescopes (LATs) on board Fermi can monitor the sky every 3 hr.

Along with the continuous monitoring of persistent sources, various missions were planned for detecting transient sources, such as gamma-ray bursts (GRBs) and/or flaring activity of active galactic nuclei (AGNs). For example, BATSE on board CGRO (Fishman et al. 1992) carried eight scintillation detectors to cover the entire sky, localizing more than 2700 GRBs in the 10 yr observation. HETE2 (Kippen et al. 2003) and Swift (Gehrels et al. 2004) are also missions dedicated to finding and providing quick localization of GRBs by means of prompt follow-up X-ray and optical/UV observations. In general, these all-sky observations provide a large amount of data; thus, it is not easy to pick up interesting events or phenomena without overlooking some. Moreover, when the exposures for specific fields are very short, detection of persistent sources and/or transient events may suffer from low photon statistics, especially when the background dominates the signal. Development of a new protocol dedicated to the analysis of all-sky monitoring is required to discriminate real and unreal events effectively and quickly without any human bias.

In this context, various analytical methods have been developed for source identification and modeling of foreground diffuse gamma-ray emission (DGE). In DGE modeling, the cosmic-ray (CR) intensities and spectra are modeled by considering their propagation in the galaxy (Ackermann et al. 2008, 2012). In contrast, Bayesian-inference-based analysis reconstructs flux contributions to the diffuse and point-like sources (Guglielmetti et al. 2009). In addition, the D3PO inference algorithm reaches de-noising, de-convoluting, and decomposing (Selig & Enßlin 2015). For variable source detection, the Fermi data are processed using automated science processing (ASP) in the LAT Instrument Science Operations Center (ISOC; Cameron 2012). ASP searches for flares by calculating the flux using maximum likelihood fitting. Apparently, this is the most accurate approach when sources on the catalog are filtered to some extent; however, developing and installing such an algorithm is challenging for individual users. In this context, aperture photometry light curves (Lenain 2017) and Fermi all-sky variability analysis (FAVA; Ackermann et al. 2013; Abdollahi et al. 2017) are simple and fast analytical methods for variable source search. These methods are convenient; however, the quality of the images is not improved.

As for the imaging analysis, significant progress has been made in machine learning (ML), especially deep learning, to find feature values in images of interest, for application in face recognition systems. In medical applications, similar approaches have already been implemented to determine lesion regions, for example, based on computed tomography images (Kalinovsky et al. 2017). An obvious advantage of using ML is that it reads images in a short time with high accuracy without any human bias in the most ideal cases. However, similar ML techniques have been implemented only very recently in astronomical research. For example, a new computer program, Morpheus, has been developed to analyze astronomical image data pixel by pixel to identify and classify all galaxies and stars (Hausen & Robertson 2020). In addition, spectral modeling of a Type II supernova based on the ML approach is being developed (Vogl et al. 2020). Application of ML techniques to all-sky survey data obtained by X-ray and gamma-ray satellite missions is also crucial, not only to best allocate resources for the data analysis but also to examine new astronomical phenomena that have yet to be explored.

In this study, we proposed a new ML approach (1) to improve the quality of low-statistic images and (2) to find variable sources and possibly even transient sources automatically from all-sky gamma-ray maps obtained with Fermi-LAT. In particular, short-observation data (1-week integration) were prepared as input, and long-observation data (1 yr integration) were predicted using three independent ML algorithms—dictionary learning (Kreutz-Delgado et al. 2003), U-net (Ronneberger et al. 2015), and Noise2Noise (Lehtinen et al. 2018)—for detailed/quantitative comparison with actual observational data. The remainder of this article is organized as follows. In Section 2, the details of the Fermi-LAT analysis and data processing before the input ML algorithms are presented. In Section 3, the architectures of the three ML algorithms are described. The results of ML predictions are presented in Section 4. In Section 5, all-sky maps (ASMs) generated by ML algorithms and the ground-truth (GT) map are compared in detail in selective regions to examine the possible origin of the mismatch. In Section 6, the conclusions are presented.

2. Observations and Analysis

2.1. Fermi-LAT Observations

In this study, ML-based algorithms were applied to a gamma-ray ASM observed by Fermi-LAT. The first objective was to fabricate a high-quality ASM with sufficient photon statistics such that the ASM uncertainties are small compared to the counting uncertainties of shorter timescale maps when only a low-quality ASM is known and/or provided as an input. As discussed in Section 3, dictionary learning and U-net require the image pairs of the short-observation map and long-observation map for image prediction. Moreover, because the Noise2Noise demands image pairs with different noise patterns, two types of short-observation maps, corresponding to the first and second week in each year, were prepared. Therefore, a long-term observation map and two types of short-term observation maps in the energy range from 100 MeV to 300 GeV were created at the beginning of each year for years 2009–2019. In this study, all energy values were incorporated into a single map. We produced the ASMs using gamma rays in the P8R3_SOURCE event selection and the associated P8R3_SOURCE_V2 instrument response functions (IRFs). In addition, to correct the ASM for exposure, we calculated an exposure map with one energy bin using gtexpcube2, which is the standard analysis tool provided with Fermi-LAT ScienceTools. We also divided the ASM by the generated exposure map. In addition, the zenith angle cut was set to 90° to remove gamma rays from Earth's limb in the analysis. In this study, ASMs were produced in the galactic coordinates in the Hammer–Aitoff projection. We generated count maps with a resolution of 0.1 deg2 per pixel.

The ML algorithms assume that all sources are persistent over 1 yr, keeping their fluxes exactly comparable to what has been observed in the first week. Thus, a large deviation of the ML maps and the GT map may suggest that variable sources are included in regions of interest (ROIs). In such cases, a light curve was generated by using binned analysis to investigate the flux variability for specific gamma-ray sources. For each source, a 15° × 15° ROI centered at its 4FGL catalog (Abdollahi et al. 2020) was chosen, and a single power-law model was applied to the data. When creating the light curve, the prefactor and the index of the central source were free, while the parameters of the other sources within the ROI were fixed. As the background model, the isotropic and diffuse backgrounds were modeled using ISO_P8R3_SOURCE_V 2_v1.txt and gll_iem_v07.fits, respectively. Moreover, only the flux bins with test static (TS) > 9 (corresponding to ∼3σしぐま) were used for each light curve.

2.2. Preprocessing for ML

The following four processing steps were applied to each ASM obtained in Section 2.1 before training the ML algorithms:

  • 1.  
    applying Gaussian blurring;
  • 2.  
    applying the natural logarithm function;
  • 3.  
    standardization of the pixel values in each map;
  • 4.  
    data augmentation.

First, Gaussian blurring was performed for each ASM to reduce the number of pixels whose value equals zero. The blurring process is necessary to prevent ML algorithms from learning only map smoothing. Subsequently, the natural logarithm function was applied to these maps owing to the high variance of pixel values; then, the pixel values were normalized from 0 to 1. Finally, random sampling was performed for data augmentation. In this process, the same regions were extracted from the ASM pairs with short-observation and long-observation maps, and the extracted images were randomly flipped. The size of the extracted images was 64 × 64 pixels, which corresponds to 6° × 6° in the galactic center. The random sampling was repeated 2000 times for each ASM pair. Detailed information about the components of the data sets is presented in Table 1.

Table 1. Components of Data Sets Prepared for Training and Testing of ML Processing

 TrainingValidationTest
Sample number1800020001
Observation year2009–201720182019
Extracted size (pixel)64 × 6464 × 643600 × 1800
Extracted size (deg)6 × 66 × 6360 × 180

Download table as:  ASCIITypeset image

Hereinafter, "input map" and "GT map" refer to the ASMs, which cover the time intervals MJD 58486.034–58493.026 (2009 January 3–9) and MJD 58486.034–58857.048 (2009 January 2–2020 January 9), respectively. The input map was blurred with a 11 pixel × 11 pixel Gaussian kernel. In the input map, standard deviation, which defines the smoothing effect, equals 3.0. Additionally, we will use "blurred map" to refer to blurred versions of the ASM, which were blurred with a 15 pixel × 15 pixel Gaussian blur kernel with a standard deviation equal to 100. "ML maps" refers to an ASM generated from the input map using three ML-based algorithms (dictionary learning, U-net, and Noise2Noise). We also note that the ML algorithm applied in this study only suggests the source variability in the corresponding image pixels. Therefore, detailed follow-up analysis such as drawing light curves of candidate sources was necessary to check the variability, as described in detail in Section 5.2.

3. Prediction Algorithms

In this study, three ML algorithms were used for prediction: dictionary learning, U-net, and Noise2Noise. The purpose and theory of these algorithms are detailed in the following sections.

3.1. Dictionary Learning

Dictionary learning is generally used for dimensional reduction in the field of signal processing (Tosic & Frossard 2011). Unlike the deep-learning algorithms, dictionary learning is based on linear algebra; hence, one can sequentially trace the obtained results using this technique. Dictionary learning involves three steps: preparing a pair of databases, creating dictionaries, and generating images. First, five patches with 16 × 16 pixels were extracted randomly from each training sample of a short-observation map. The patches were arranged in columns and aligned to create a database of short-observation data (Ashort). In the same manner as Ashort, a database of long-observation data (Along) was also created. Then, the arrays, called the "dictionary," were calculated. In this study, the dictionaries from short-observation data (Dshort) and long-observation data (Dlong) were created from Ashort and Along by sparse representation. After the dictionaries were calculated, the test samples from the input map were represented by Dshort and sparse matrix C. Finally, Dshort was replaced with Dlong to generate images corresponding to the GT map. The details of the calculation in dictionary learning are provided elsewhere (Sato et al. 2020). In this study, scikit-learn (version 0.21.2), which is a library of ML algorithms used in Python, was used for dictionary learning and sparse coding.

3.2. U-net Architecture

Originally, a U-net was developed for masking the medical images (Dong et al. 2017; Han & Ye 2018; Zhou et al. 2018). An example of the algorithm architecture is illustrated in Figure 1. U-net consists of two paths: the contracting path and expanding path. Two convolution layers and a max pooling layer are repeated four times in the contracting path. In this study, 3 × 3 convolution was used with leaky rectified linear unit (ReLU) activation and batch normalization. The number of channels was doubled for each downsampling. The expanding path consists of two convolution layers and a deconvolution layer with 3 × 3 filter sizes. The ReLU function was used as the activation function after the convolution layers except the last one. After the last convolutional layer, a hyperbolic tangent function was used. In this study, the "GeForce 1080Ti" graphics processing unit (GPU) processer was employed, and the U-net was implemented with TensorFlow (version 2.1.0) and Keras (version 2.3.1), which are libraries of deep-learning algorithms used in Python.

Figure 1.

Figure 1. U-Net architecture with two downsampling layers. Black arrows denote shortcut path.

Standard image High-resolution image

3.3. Noise2Noise Architecture

Noise2Noise, which is a semisupervised learning algorithm, is used to reconstruct clean signals from corrupted ones (Wu et al. 2019; Hasan et al. 2021; Díaz Baso et al. 2019). This algorithm can perform signal de-noising without the need for clean signal data sets. Therefore, Noise2Noise is applicable to cases where clean signals are difficult to obtain. The superresolution residual network (SRResNet; Ledig et al. 2017) used in this study consists of a convolution layer with parametric rectified linear unit (PReLU) activation, six layers of residual block, two layers of convolution layer with batch normalization, and a concatenated layer, as depicted in the left panel of Figure 2. The right panel of Figure 2 depicts the details of the residual block. The residual block consists of a convolution layer, batch normalization, and PReLU activation. In this study, the filter number and size in the convolution layer were set as 64 and 3 × 3, respectively. The software and GPU processer were used in the same way as for U-net in Section 3.2.

Figure 2.

Figure 2. ML algorithm architecture for Noise2Noise; the architecture of SRResNet (left) and the residual block (right) is shown.

Standard image High-resolution image

4. Results

4.1. Improvement of Image Quality

To confirm the improvement of maps in detail, close-ups of the various regions in Figure 3 were extracted, as shown in Figure 4. Image noise resulting from low photon statistics is reduced by ML algorithms; thus, the parts of the point source are enhanced. Moreover, the feature of the diffuse gas emission around the galactic plane is also well reproduced.

Figure 3.

Figure 3. Results of applying ML algorithms to ASM obtained with Fermi-LAT. Top: input map, blurred map, and GT map; middle: ML maps generated by each ML algorithm (dictionary learning, U-net, and Noise2Noise); bottom: relative-ratio maps, which were calculated by dividing the GT map by each ML map.

Standard image High-resolution image
Figure 4.

Figure 4. Extracted part of the input map, blurred map, and GT map are shown (top), and the same parts of ML maps are displayed (bottom).

Standard image High-resolution image

To evaluate the image improvement in ML maps, we applied a root mean square error (RMSE), which is the standard evaluation index in computer vision. The RMSE represents the squared error of each pixel value in the two images and was calculated using the following equation:

Equation (1)

where n and Vi denote the number of image pixels and the ith pixel value in each input image, respectively. An RMSE that is closer to zero represents a smaller error between the input images. Similarly, we also computed the structural similarity (SSIM) between the produced maps and the GT map. The SSIM, which represents visual similarity, was calculated using

Equation (2)

where μみゅー and σしぐま denote the average value and standard deviation of the image, respectively. C1 and C2 are small constants for avoiding instability in the case in which ${\mu }_{x}^{2}+{\mu }_{y}^{2}$ or ${\sigma }_{x}^{2}+{\sigma }_{y}^{2}$ is close to zero. Details of the constants are described in Wang et al. (2004). Figure 5 depicts the results of RMSE and SSIM evaluation. In Figure 5, the horizontal axis represents the observation time, and the vertical axis represents the RMSE or SSIM values. We discuss the evaluation results in Section 5.3.

Figure 5.

Figure 5. Evaluation results in RMSE (top) and SSIM (bottom) based on the GT map (1 yr map). Maps with observation periods of 1 week, 4 weeks, 8 weeks, 12 weeks, and 16 weeks were evaluated and plotted in blue. The result of applying ML maps is also shown.

Standard image High-resolution image

4.2. Comparing ML Maps to the GT Map

For source detection, we compared the excess of pixel values in the ML maps corresponding to the source position, as already listed/known in the 4FGL catalog. The bottom panel of Figure 6 shows an example of a projection histogram, the region of which is shown in the top panel. In Figure 6, 15 representative sources with the highest average flux are indicated. The signal-to-noise ratio (S/N) in the projection was improved by ML algorithms. However, unlike input and GT maps based on photon statistics, pixel values in the ML maps lose any physical meaning; thus, we could not evaluate TS, which is often applied in the analysis of Fermi-LAT to evaluate the significance of source detection. We will develop such advanced algorithms for ML maps in the future; however, it is beyond the scope of this paper.

Figure 6.

Figure 6. Example of 1D projection histogram from the input map, U-net map, and GT map. The extracted region (top) and 1D histogram (bottom) are shown. Representative sources in the 4FGL catalog whose average flux is larger than 4 × 10−9 are indicated.

Standard image High-resolution image

To find the correspondence between the ML maps and the GT map, the 1D projections along the same galactic latitude l were compared for various regions. An example of a projection histogram obtained by the GT map, U-net map, and Noise2Noise map is depicted in the top panel of Figure 7. The extracted part is the same as in Figure 6. The projection was created by integrating the pixel values for 15° in the direction of galactic latitude l. For a quantitative comparison, the values in the ML maps were divided by those in the GT map. The bottom panel of Figure 7 indicates the ratio of the ML maps to the GT map. Three regions with large differences are marked by orange circles. Figure 8 shows close-ups around region 2 to illustrate such a difference. While no evident sources exist in the input map and ML maps, a bright point source, which is identified as PKS 2052−47, is found only in the GT map. In the opposite case shown in Figure 7, there exist regions in which the ratio is smaller than unity, as represented by region 4 in Figure 9. Figure 10 shows close-ups around region 4. The details of regions with apparent "mismatch" are listed in Table 2.

Figure 7.

Figure 7. Example of projection histogram and division: the 1D projection (top) and the division values obtained by dividing the GT map by the values of ML maps (bottom) are shown. Orange circles show the large difference between the 1D projection calculated from the ML maps and the projection from the GT map.

Standard image High-resolution image
Figure 8.

Figure 8. Corresponding part of large division value (PKS 2052−47): the input map, blurred map, and GT map are shown (top), the same parts of the ML maps are displayed (bottom), and the part with a large difference is surrounded by an orange circle.

Standard image High-resolution image
Figure 9.

Figure 9. Negative case of projection histogram and division: the extracted part (top) and the 1D projection (middle) are shown, the division values obtained by dividing the GT map by the values of the ML maps are shown (bottom), and orange circles show the large difference between the 1D projection calculated from ML maps and the projection from the GT map.

Standard image High-resolution image
Figure 10.

Figure 10. Corresponding part of large division value (PKS 1824−582): the input map, blurred map, and GT map are shown (top), the same parts of the ML maps are displayed (bottom), and the part with a large difference is surrounded by an orange circle.

Standard image High-resolution image

Table 2. List of Source IDs and Positions Corresponding to the Circled Part in Figures 7 and 9

Regionassoc4FGLSource NameSource TypeRedshiftR.A.Decl.
     J2000J2000
Region 13C 454.3J2253.9+1609AGN/FSRQ0.859343.490616.1482
Region 2PKS 2052−47J2056.2−4714AGN/FSRQ1.489314.0715−47.2369
Region 3PKS 0454−234J0522.9−3628AGN/FSRQ1.00374.2608−23.4145
Region 4PKS 1824−582J1829.2−5813AGN/FSRQ1.531277.3108−58.2323

Download table as:  ASCIITypeset image

5. Discussion

5.1. Comparing Various ML Algorithms

In this study, three types of ML algorithms were applied: dictionary learning, U-net, and Noise2Noise. Dictionary learning is based on linear algebra and can be used to trace the calculation process sequentially. However, dictionary learning has a low representation capability compared with other deep-learning algorithms and reflects a relatively low quality image compared with other algorithms presented in Figure 4. Thus, dictionary learning is less effective than other deep-learning algorithms, for the purpose of this study. In the U-net architecture, the input and output layers are connected by two paths: an extended path and a shortcut connection. Since these two paths are considered to capture detailed features and context, U-net can successfully predict the long-observation map from the input map with high precision (see Figure 4). However, in Noise2Noise, the algorithm captures features of statistical noise by comparing two types of short-observation maps. Therefore, the algorithm is used to generate the ASM by neglecting the statistical noise, which can sometimes overestimate the significance of each point source. Thus, U-net may be superior to Noise2Noise for predicting a long-observation map. In this study, U-net was considered to be the most reliable/precise for predicting the long-observation map. Furthermore, because these three ML algorithms are based on completely different theories, ML maps are more reliable if the outputs from each ML algorithm are similar to each other.

5.2. Origin of Mismatch between the GT Map and ML Maps

In this section, a possible reason for the mismatch between the GT map and ML maps, as shown in Figures 7 and 9, is discussed. First, the regions of the largest mismatch, indicated as orange circles in Figures 7 and 9, are discussed. The ML maps are generated based on the hypothesis that the sources are persistent over 1 yr, which is, however, not the case for variable and/or transient sources. Thus, the observed mismatch between the ML maps and GT map is likely associated with the variable and transient sources, such as the flaring activity of AGNs. To confirm this assumption, the sources listed in Table 2 are analyzed in more detail with the Fermi analysis tool, as described in Section 2.1. The light curves around each source are shown in Figure 11. For example, a variable source in region 1 corresponds to a flat spectrum radio quasar (FSRQ) 3C 454.3 with a redshift of z = 0.859 (Table 2). As shown in Figure 11, the flux of 3C 454.3 in the first week was much smaller than the average flux, shown as an orange line, integrated over a year. Similarly, regions 2 and 3 were identified as two FSRQs: PKS 2052−47 and PKS 0454−234, respectively. Contrary to the sources in which flux substantially increased after the first week, the average flux over the year is smaller than the first week's flux in the case of region 4, which corresponds to FSRQ PKS 1824−582. Obviously, such temporal variability in fluxes explains the observed mismatch between the ML maps and the GT map, as shown in Figures 7 and 9. As can be seen in this figure, point sources with increasing photon flux appear to be brighter in the bottom panel of Figure 3. On the contrary, point sources with decreasing flux appear faint.

Figure 11.

Figure 11. Fermi-LAT light curve and variation of photon index for each source shown in Table 2 over the time interval from 2019 January to 2019 December in 1-week time bins. In each graph, only flux bins with TS > 9 are plotted. Because PKS 2052−47 had too few photons and the TS did not exceed 9, the data for the first 2 months from the beginning of 2019 were used as the first bin. Similarly, the light curve for PKS 1824−582 has few bins with TS > 9 throughout the year, so its light-curve data were binned with 1-month intervals, while only the first bin was binned with 1-week data because the number of photons was sufficient for the first week. Top panel: changes in the 100 MeV < E < 300 GeV flux. The orange line shows the initial 1-week flux. Middle panel: changes in the power-law photon index. Bottom panel: changes in the TS.

Standard image High-resolution image

However, there is a systematic difference between the GT map and ML maps around the galactic longitude 270° < l < 330° in the bottom panel of Figure 7, which cannot be explained by either the statistical fluctuation or a combination of variable sources. We also confirm the difference in the bottom panel of Figure 3. Such a systematic gradient is thought to be caused by the nonuniformity of the exposures inherent in the 1-week observation, which were used as input for ML algorithms. Although exposure map correction is performed based on the position and angle of the satellite, such nonuniformity affects the reconstructed counting map, especially in the case of short observation times. In fact, we confirmed that such a gradient is reduced if 2-week ASM maps are used as input, but the optimization of the time duration for the input maps is beyond the scope of this study. Additionally, ASMs in this paper were produced with no energy binning; hence, the exposure map might not correct the actual exposure. This is because the accuracy of integrating the exposure over the energy band depends on the smoothness of the source spectrum across the energy band, and these are quite large energy bands.

5.3. Quality of ML Maps

As shown in Figure 5, the ML-based algorithms were effective in improving the ASM quality. In particular, U-net was the best among the three ML algorithms. The RMSE and SSIM assessments of the U-net map were equivalent to approximately 10 and 12 weeks, respectively. Considering that the exposure map and variable source were incorrect, the image quality was improved by more than 10 weeks. Despite the significant improvement, faint source detection was considered difficult for our algorithms owing to image blur. Some peaks around the source location were blurred, as shown in Figure 6. In general, for ML processing, the loss function must be defined. "Training ML algorithms" refer to parameter optimization, which could be used for minimizing the loss function. The RMSE, which is one of the standard loss functions, was used in this study. However, in the case of using MSE or RMSE, it is known that high-frequency components are difficult to reproduce. Due to the difficulty of point-source detection, some schemes are necessary to improve the detection efficiency. For example, adopting the sum of point-source detectability and RMSE as the loss function may improve the performance. Other approaches are mentioned in Section 5.5.

5.4. Comparison with Existing Method

Compared with other techniques described in Section 1, our ML-based method has the following three advantages:

  • 1.  
    The ML algorithm can be easily implemented by end users. Moreover, any assumption or modeling of foreground diffuse emission is unnecessary. Owing to its simplicity, the ML ASM can be obtained within less than 10 minutes.
  • 2.  
    S/N on the image is improved in terms of both RMSE and SSIM, although the optimum algorithms to detect faint point sources remain to be determined in a future work.
  • 3.  
    By comparing the ML and GT maps, variable sources can be detected automatically in ASM in either case (increasing and/or decreasing) of fluxes.

Conversely, the ML method generally loses the physical properties of the source images, such as the energy of each photon and the statistical significance. Thus, one should consider such an ML-based image as a starting point for finding possible interesting sources. These sources would be chosen for subsequent detailed analysis based on the standard analysis tool released by the Fermi-LAT collaboration. Detailed modeling of DGE, as well as evaluation of the TS, is essential even after the application of ML methods presented in this paper.

In this context, the D3PO inference algorithm, which is one of the latest modeling techniques, demonstrated excellent noise reduction. However, analysis with complex models is required, which is labor-intensive and computationally expensive. Our ML-based method, on the contrary, is model independent and can be completed by simple implementation, as mentioned above. While the method is cost-effective, ML maps lose physical properties, such as energy spectrum.

Finally, we discussed the variable source detection in comparison to existing methods. For source variability in the 4FGL catalog, the variability index, such as TSvar (Nolan et al. 2012), was calculated from light curves using 1 yr bin data. For bright and variable sources, light curves are automatically updated daily for cataloged or well-known sources. In contrast, our ML methods have the potential to automatically detect serendipitous variable sources. However, as mentioned earlier, our ML-based technique loses any physical properties of the sources; hence, a detailed analysis is necessary. In this context, FAVA, which is a simple and effective method for detecting variable sources, was employed. However, as FAVA detects the deviation between the map expected from long-term observations and the one from short-term observations, the S/N does not change. Our algorithms improved the S/N around the point source; hence, the variable source detection has the potential to achieve higher accuracy than FAVA.

5.5. Future Prospects

In this study, an initial 1-week map obtained with Fermi-LAT in 2019 was prepared as a test sample. However, the prediction method is applicable not only to gamma-ray sky maps but also to other maps obtained in different energy bands, such as radio, optical, and X-ray ASMs. Moreover, one can set any time interval and duration depending on the research purpose, astrophysical phenomenon, and source. For example, one can use a timescale that is as short as minutes to detect various transient sources, such as GRBs, without human bias at any point by applying ML algorithms. Longer timescales of weeks to months are more favorable for searching tidal disruption events (Burrows et al. 2011; Gezari et al. 2012) and/or supernovae.

In this research, a 1-week map was used as an input map to demonstrate the concept and performance of various ML algorithms as applied to Fermi-LAT data. In planned future work, an attempt will be made to find transient sources by setting ASMs generated at different timescales, to widely cover scales of minutes to days as input. Moreover, another interesting possibility is that one can predict longer observation maps if the observations are continued, even after the mission has been completed. For example, if a 1 yr ASM is used as the input, it may be possible to predict a 100 yr ASM from existing data. Clearly, such predictions may or may not be correct, as shown in Section 5; however, it can provide strong motivation for future missions and important scientific aspects to be explored in the future. Similar ML approaches have been proposed to confirm the giant but weak emission structure of diffuse gas, such as the Fermi bubble (Su et al. 2010) and NPS/Loop I (Kataoka et al. 2018), which was proposed using Gaussian mixture algorithms (Müller et al. 2018).

For future research directions, three different approaches are promising. The first is to train the ML algorithm separately for the point-source and DGE regions. In this study, although we used ASM data for training, the DGE and point-source regions were separated to train the ML algorithms. The second approach is to apply other ML algorithms, especially deep-learning algorithms. In this study, U-net and Noise2Noise were considered to be effective for improving the ASM obtained by Fermi-LAT. However, it is necessary to search other ML algorithms, such as the generative adversarial network (GAN) type or Fourier convolutional neural network (FCNN) type, for more practical applications. By using the FCNN-type algorithms, it may be possible to obtain the features represented in the frequency domain. Another approach is to use a large amount of data for training. In this study, only actual observational data were prepared. The amount of training data can be easily increased using simulation data.

6. Conclusion

A novel application of ML algorithms to analyze ASMs obtained with the Fermi-LAT was demonstrated in this paper. An ASM corresponding to a long-observation map was successfully generated from a short-observation map by using three types of ML algorithms. ML-based methods make it possible to improve image quality, not only for point sources but also for diffuse gas structures. Furthermore, these methods are useful for detecting variable sources automatically without human bias by analyzing large deviations between the simulated ML map and the GT map. In addition, the method can be extended to other satellite data, even if the satellite has completed its observations. Thus, in future work, the proposed method will be applied to other satellite data, such as eROSITA on SRG and/or BATSE on board CGRO.

We thank an anonymous referee for their careful advice for improving the manuscript. This research was supported by the Japan Society for the Promotion of Science (JSPS) KAKENHI (grant No. 20H00669).

Please wait… references are loading.
10.3847/1538-4357/abf48f