This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Learning for Line Intensity Mapping Observations: Information Extraction from Noisy Maps

Kana Moriwaki Department of Physics, The University of Tokyo, 7-3-1 Hongo, Bunkyo, Tokyo 113-0033, Japan Masato Shirasaki National Astronomical Observatory of Japan (NAOJ), Mitaka, Tokyo 181-8588, Japan The Institute of Statistical Mathematics, Tachikawa, Tokyo 190-8562, Japan Naoki Yoshida Department of Physics, The University of Tokyo, 7-3-1 Hongo, Bunkyo, Tokyo 113-0033, Japan Kavli Institute for the Physics and Mathematics of the Universe (WPI), UT Institutes for Advanced Study, The University of Tokyo, 5-1-5 Kashiwanoha, Kashiwa, Chiba 277-8583, Japan Institute for Physics of Intelligence, School of Science, The University of Tokyo, 7-3-1 Hongo, Bunkyo, Tokyo 113-0033, Japan
Abstract

Line intensity mapping (LIM) is a promising observational method to probe large-scale fluctuations of line emission from distant galaxies. Data from wide-field LIM observations allow us to study the large-scale structure of the universe as well as galaxy populations and their evolution. A serious problem with LIM is contamination by foreground/background sources and various noise contributions. We develop conditional generative adversarial networks (cGANs) that extract designated signals and information from noisy maps. We train the cGANs using 30,000 mock observation maps with assuming a Gaussian noise matched to the expected noise level of NASA’s SPHEREx mission. The trained cGANs successfully reconstruct Hα\rm H\alpha emission from galaxies at a target redshift from observed, noisy intensity maps. Intensity peaks with heights greater than 3.5σnoise3.5~{}\sigma_{\rm noise} are located with 60 % precision. The one-point probability distribution and the power spectrum are accurately recovered even in the noise-dominated regime. However, the overall reconstruction performance depends on the pixel size and on the survey volume assumed for the training data. It is necessary to generate training mock data with a sufficiently large volume in order to reconstruct the intensity power spectrum at large angular scales. The suitably trained cGANs perform robustly against variations of the galaxy line emission model. Our deep-learning approach can be readily applied to observational data with line confusion and with noise.

high-redshift galaxies — large-scale structure of the universe — observational cosmology
journal: ApJLsoftware: Cloudy (Ferland et al., 2013), Pinocchio (Monaco et al., 2013)

1 Introduction

The large-scale structure of the universe contains rich information on galaxy formation and on the nature of dark matter and dark energy. Line intensity mapping (LIM) is an emerging observational technique that measures the fluctuations of line emission from galaxies and intergalactic medium. With typically low angular and spectral resolutions, LIM can survey an extremely large volume. Future LIM observations are aimed at detecting emission lines at various wavelengths: Hi 21cm line (e.g., SKA; Koopmans et al., 2015), FIR/submillimeter lines such as [Cii] and CO (e.g., TIME; Crites et al., 2014), and ultraviolet/optical lines such as Lyα\rm\alpha and Hα\rm H\alpha (e.g., SPHEREx; Doré et al., 2014).

While LIM has the advantage of being able to detect all contributions including emission from faint, dwarf galaxies, there is a serious contamination problem, the so-called line confusion. Because individual line sources are not resolved in LIM observations, foreground/background contamination cannot be easily removed. So far, only a few practical methods have been proposed to extract designated signals. Statistics-based approaches include cross-correlation analysis with galaxies/emission sources from the same redshift (e.g., Visbal & Loeb, 2010), and one that utilizes the anisotropic power spectrum shape (e.g., Cheng et al., 2016). Cheng et al. (2020) devise a method based on sparsity modeling that successfully reconstructs the positions and the line luminosity functions of point sources from multifrequency data.

Earlier in Moriwaki et al. (2020), we have proposed a deep-learning approach to solve the line confusion problem. We use conditional generative adversarial networks (cGANs), which are known to apply to a broad range of image-to-image translation problems. Our cGANs learn the clustering features of multiple emission sources and are trained to separate signals from different redshifts. It is shown that deep learning offers a promising analysis method of data from LIM observations. However, in practice, various noise sources can cause a serious problem. Faint emission-line signals from distant galaxies are likely overwhelmed by noise even with the typical level of next-generation observations.

In this Letter, we propose to use cGANs to effectively de-noise line intensity maps. We show that suitably trained cGANs successfully reconstruct the emission line signals on a map, and recovers basic statistics of the intensity distribution. All such information extracted from noisy maps can be used for studies on cosmology and galaxy population evolution.

2 Methods

We consider Hα\rm H\alpha emission from galaxies at z=1.3z=1.3 and observed at 1.5 μm\rm\mu m. The Hα\rm H\alpha emission is one of the major target lines of future satellite missions such as SPHEREx (Doré et al., 2014) and CDIM (Cooray et al., 2019). We develop cGANs that extract Hα\rm H\alpha signals from noisy observational data. We first describe how we generate mock intensity maps that are used for training and test. We then explain the basic architecture of our cGANs. Further technical details can be found in Moriwaki et al. (2020).

2.1 Training and test data

We prepare a large set of training and test data. We use a fast halo population code PINOCCHIO (Monaco et al., 2013) that populates a cosmological volume with dark matter halos in a consistent manner with the underlying linear density field. We generate 300 (1000) independent halo catalogs with a cubic box of 280h1280h^{-1} Mpc on a side for training (test)111In the rest of this Letter, we adopt Λ\rm\LambdaCDM cosmology with Ωm=0.32\Omega_{m}=0.32, ΩΛ=0.68\Omega_{\Lambda}=0.68, and h=0.67h=0.67 (Planck Collaboration VI, 2018).. The smallest halo mass is 3×1010M3\times 10^{10}~{}\rm M_{\odot}. We then assign Hα\rm H\alpha luminosities to the individual halos to obtain a three-dimensional emissivity field. The halo mass-to-luminosity relation is derived using the result of a hydrodynamics simulation Illustris-TNG (Nelson et al., 2019). We assume that the line luminosity is given by a function of the star-formation rate of the simulated galaxy as

Lline=10Aline/2.5ClineM˙,\displaystyle L_{\rm line}=10^{-A_{\rm line}/2.5}C_{\rm line}\,\dot{M}_{*}, (1)

where we adopt Aline=1.0A_{\rm line}=1.0 mag, and ClineC_{\rm line} is a coefficient table computed using the photoionization simulation code Cloudy (Ferland et al., 2017).

We work with two-dimensional images (intensity maps) in order to make the best use of modern image translation methods, although, in principle, it is possible to construct neural networks that read and generate three-dimensional data (e.g., Zhang et al., 2019). We generate two-dimensional Hα\rm H\alpha intensity maps by projecting the three-dimensional emissivity fields along one direction.

For each realization of the training (test) data, 100 maps (1 map) with an area of (0.85deg)2(0.85~{}\rm deg)^{2} are generated by projecting random portions of an emissivity field. A total of 30,000 training data and 1000 test data are generated in this manner. The intensity maps are pixelized with the angular and spectral resolution of SPHEREx listed in Table 1.

Finally, realistic mock observation maps are generated by adding Gaussian noise to the Hα\rm H\alpha intensity map. We adopt the noise level of ”SPHEREx deep” whose 5σn5~{}\sigma_{\rm n} sensitivity per pixel at λ=1.5μ\lambda=1.5\rm\mum is 22 mag, corresponding to σn=2.6×106erg/s/cm2/sr\sigma_{\rm n}=2.6\times 10^{-6}~{}\rm erg/s/cm^{2}/sr. The maps are normalized by 1.0×104erg/s/cm2/sr1.0\times 10^{-4}~{}\rm erg/s/cm^{2}/sr before input to the networks.

Table 1: Observational parameters of SPHEREx deep (Doré et al., 2014).
Field of view 200deg2200~{}\rm deg^{2}
Angular resolution ll 6".2×6".26".2\times 6".2
Spectral resolutiona RR 41.5
Sensitivitya σn\sigma_{\rm n} [erg/s/cm2/sr][\rm erg/s/cm^{2}/sr] 2.6×1062.6\times 10^{-6}
Refer to caption
Figure 1: From left to right, we show the observed map (Hα\rm H\alpha + noise), the true Hα\rm H\alpha map, and the reconstructed Hα\rm H\alpha map. Each map has 0.85 deg on a side. The maps are smoothed with a Gaussian beam with σ=0.6\sigma=0.6 arcmin for better visibility. The intensities are in units of 107ergs1cm2sr1\rm 10^{-7}~{}erg{\rm s}^{-1}{\rm cm}^{-2}{\rm sr}^{-1}. Note that the observed map is noise-dominated. See the color-bars on the right.

2.2 Network architecture

We develop cGANs using the publicly available pix2pix code (Isola et al., 2016). The cGANs consist of two adversarial convolutional networks: a generator and a discriminator. The generator, consisting of 8 convolutional and 8 de-convolutional layers, outputs a map GG (”reconstructed map”) from an observed map xx. The discriminator, consisting of 4 convolutional layers, returns a value DD for the input of (x,y)(x,y) or (x,G[x])(x,G[x]) with yy denoting the Hα\rm H\alpha map. The value DD indicates the probability that the input is not (x,G[x])(x,G[x]) but (x,y)(x,y). During the training, the two networks are updated repeatedly in an adversarial way; the generator is updated so that it deceives the discriminator (i.e., D(G[x])D(G[x]) should get closer to 1), while the discriminator is updated so that it gets better accuracy (i.e., D(x,y)D(x,y) and D(x,G[x])D(x,G[x]) get closer to 1 and 0, respectively).

Specifically, the parameters in the generator (discriminator) are updated to decrease (increase) the loss function

=cGAN(G,D)+λL1(G),\mathcal{L}=\mathcal{L}_{\rm cGAN}(G,D)+\lambda\mathcal{L}_{\rm L1}(G), (2)

where

cGAN(G,D)\displaystyle\mathcal{L}_{\rm cGAN}(G,D) =\displaystyle= logD(x,y)+log[1D(x,G[x])],\displaystyle\log D(x,y)+\log[1-D(x,G[x])], (3)
L1\displaystyle\mathcal{L}_{\rm L1} =\displaystyle= 1Npix|yG(x)|.\displaystyle\frac{1}{N_{\rm pix}}\sum|y-G(x)|. (4)

Note that we include an additional term L1\mathcal{L}_{\rm L1} that is known to ensure better performance by imposing the condition that the values of the corresponding pixels in the true and reconstructed maps should be close (Isola et al., 2016). In each round of training, the loss function is computed with a mini-batch. After some experiments, we set λ=1000\lambda=1000 and batch size 4. The networks are trained for 8 epochs. We adopt these parameter values throughout the present study.

3 Results

Figure 1 shows the reconstruction performance of our cGANs. Our networks reduce the noise and successfully extract the true Hα\rm H\alpha signals. It is remarkable that both the source positions and the intensities are reproduced well, even though the observed map is noise dominated. A simpler approach in such a noise-dominated case would be to select only high signal sources in an observed map. However, we find that, if we select pixels with signals greater than 3.5σn3.5\sigma_{\rm n} from the observed maps, only 20 % of them are true sources (see also Figure 2). With our networks, about 60 % of the reconstructed pixels with intensities greater than 3.5σn3.5\sigma_{\rm n} are real sources. Hence our method significantly outperforms the simple signal selection based on the local intensity.

3.1 Probability distribution function

The probability distribution function (PDF) of line intensity is an excellent statistic that can constrain galaxy populations and their physical properties (e.g., Breysse et al., 2017). We test whether our networks also recover the PDF accurately.

We first note that, in general, a single set of networks do not reproduce pixel statistics of images/maps robustly. We thus resort to training multiple networks and take the mean of the statistics reconstructed by the ensemble of networks. This technique, called ”bagging,” is known to reduce generalization errors (Goodfellow et al., 2016), and has been applied to, for instance, de-noising weak lensing convergence maps (Shirasaki et al., 2019). In practice, we average the PDFs reconstructed by 5 networks that are trained with different datasets.

Figure 2 compares the PDFs of true and reconstructed maps. The vertical dashed line indicates the 1σ1\sigma noise level. Our cGANs are able to reconstruct the PDF of the Hα\rm H\alpha intensity above 1σ\sigma. Note also that the scatter of the averaged PDF lies within the intrinsic scatter of the true Hα\rm H\alpha maps, i.e., within the so-called cosmic variance. Apparently, the networks tend to reconstruct the PDFs close to the average. This is simply caused by the bagging procedure. We have checked and confirmed that the variance of reconstructed PDFs by a single network is as large as the intrinsic one.

Refer to caption
Figure 2: The probability distribution function of Hα\rm H\alpha (red) and noise (black) maps. The shaded regions and the error bars indicate 1σ\sigma variances of the true and the reconstructed maps evaluated using 1000 test data. The PDF of a reconstructed map is computed by taking an average of PDFs of 5 reconstructed maps with 5 different networks.
Refer to caption
Figure 3: The power spectra of Hα\rm H\alpha maps. We show the result of 0.85 deg map (red) and 8.5 deg map (purple). The shaded regions and error bars correspond to 1σ\sigma variances of the true and reconstructed maps for 1000 test data. The reconstructed power spectrum is computed by taking an average of those of 5 reconstructed maps by 5 different networks. The gray (0.85 deg) and black (8.5 deg) errorbars show the noise power spectra PnP_{n} and their uncertainties Pn/NkP_{n}/\sqrt{N_{k}}.
Refer to caption
Figure 4: The leftmost panel is the one of the true Hα\rm H\alpha maps of original test dataset. The second left and middle panel show the reconstructed Hα\rm H\alpha maps for test data in which the Hα\rm H\alpha intensities are scaled with cHα=0.5c_{\rm H\alpha}=0.5, and 2.0, respectively. The color bars are scaled so that their corresponding ”true” images look the same as the leftmost image. The two panels on the right are the true and reconstructed maps when we adopt the Hα\rm H\alpha model of Silva et al. (2018). Each map has 0.85 deg on a side. All the maps are smoothed with a Gaussian beam with σ=0.6\sigma=0.6 arcmin for better visibility. The intensities are in units of 107ergs1cm2sr1\rm 10^{-7}~{}erg{\rm s}^{-1}{\rm cm}^{-2}{\rm sr}^{-1}.

3.2 Power spectrum

We further examine the ability of our cGANs to reconstruct the intensity power spectrum. To this end, we again adopt the bagging of 5 networks that are trained with different datasets. The red points with error bars and the shaded regions in Figure 3 show the power spectra of the reconstructed and true Hα\rm H\alpha maps, respectively. We also show the noise power spectrum, Pn(k)P_{n}(k), and its variance, Pn(k)/NkP_{n}(k)/\sqrt{N_{k}}, where NkN_{k} is the number of modes in kΔk/2<|𝒌|k+Δk/2k-\Delta k/2<|\bm{k}|\leq k+\Delta k/2. We adopt Δlogk=Δk/k=0.2\Delta\log k=\Delta k/k=0.2.

We notice that the reconstructed power spectra on large scales (k0.5k\lesssim 0.5) are systematically underestimated. This might be owing to the finite box size of the training data. To examine this, we train the cGANs with mock intensity maps with a larger area. For this test, we generate halo catalogs in a cubic volume of 700h1700h^{-1} Mpc on a side (see Section 2.1). Then the smallest halos populated is degraded to 3×1011M3\times 10^{11}~{}\rm M_{\odot}, but we have confirmed that the mean Hα\rm H\alpha intensity (or the total luminosity density) is not significantly different from that with our default box size of 280h1280h^{-1} Mpc. We set the side length of the pixel l=2.0arcminl^{\prime}=2.0~{}\rm arcmin and the spectral resolution R=41.5R^{\prime}=41.5. Each map has a ten times larger area of (8.5deg)2(8.5~{}\rm deg)^{2}. The noise level scales with the angular and spectral resolution as

σn=σn(ll)2RR,\displaystyle\sigma_{\rm n}^{\prime}=\sigma_{\rm n}\sqrt{\Big{(}\frac{l}{l^{\prime}}\Big{)}^{2}\frac{R^{\prime}}{R}}, (5)

where ll, RR, and σn\sigma_{\rm n} are the original angular and spectral resolution and the noise level of SPHEREx (Table 1). The resulting noise level of the wide maps is σn=1.3×107erg/s/cm2/sr\sigma_{\rm n}^{\prime}=1.3\times 10^{-7}~{}\rm erg/s/cm^{2}/sr. We adopt the same hyperparameters in the cGANs as in our default case except we set λ=200\lambda=200 and the normalization factor 1.0×106erg/s/cm2/sr1.0\times 10^{-6}~{}\rm erg/s/cm^{2}/sr for the low-resolution maps. The purple dots with error bars in Figure 3 show the power spectrum of the reconstructed wide maps. The light-pink shading indicates the 1σ\sigma dispersion of the true Hα\rm H\alpha power spectra. Clearly, the large-scale (low-kk) power spectrum is reconstructed more accurately compared to our default case. We note that the cGANs trained with the wider maps do not resolve point sources, but the peaks and voids in the reconstructed map correspond closely to the positions of groups/clusters and void regions.

Ideally, networks trained with intensity maps that have fine pixels and a large box size would be able to reconstruct both the positions of point sources (galaxies) and their large-scale clustering. Unfortunately, it becomes computationally more expensive if we set a larger number of pixels. The computational time for training roughly scales with the number of pixels, and the necessary number of training epochs could also increase. We thus suggest that one should generate training data depending on the purpose. In order to detect point sources robustly, one needs to train the networks using fine-pixel maps. If the primary purpose is to reconstruct the large-scale power spectrum, for cosmology studies for instance, then one needs to generate maps with a sufficiently large area (volume) but with coarse pixels. The reconstructed power spectra shown in Figure 3 suggest that one should adopt at least a several times larger area for training than the actual size of observed maps.

3.3 Line emission models

Refer to caption
Figure 5: The average of the probability distribution function (left) and the power spectrum (right) of the true (dashed) and the reconstructed (solid) Hα\rm H\alpha maps over 1000 test dataset. The reconstructed statistics are computed by taking means of those of 5 different networks. The cyan, red, purple, and yellow lines correspond to the results with cHα=0.5,1c_{\rm H\alpha}=0.5,1, 2, and Silva et al. (2018) model, respectively.

The line intensities of high-redshift galaxies are not well constrained observationally, and theoretical models of galaxy formation and evolution remain uncertain. Thus it is important to examine whether our GAN-based method can be applied robustly to data that are generated with different line emission models. To this end, we generate three additional sets of 1000 test data that have the same noise level but have different Hα\rm H\alpha line intensities. Two of the new test datasets are generated by simply multiplying the original Hα\rm H\alpha maps by a constant value, cHα=0.5c_{\rm H\alpha}=0.5 and 2, as

Iobs(𝒙)=cHαIHα(𝒙)+Inoise(𝒙).\displaystyle I_{\rm obs}(\bm{x})=c_{\rm H\alpha}I_{\rm H\alpha}(\bm{x})+I_{\rm noise}(\bm{x}). (6)

These cases serve as a test of the networks’ reconstruction performance against the variation in the mean intensity. We also adopt a completely different Hα\rm H\alpha model of emission, Silva et al. (2018), in which the SFR-halo mass relation derived by Guo et al. (2013) is used to assign the Hα\rm H\alpha luminosities (\propto SFR). The model reproduces the mean Hα\rm H\alpha intensity consistent with observations (Sobral et al., 2013, 2015).

Figure 4 shows examples of reconstructed images. In all the cases, the locations and amplitudes of bright peaks are reconstructed well, even though the mean intensity levels are quite different from the original data (see the color bars). It appears that the networks practically learn the noise properties and extract the signal robustly, even if the signal amplitude is different from the training data. It is remarkable that, in the cases with cHα=2c_{\rm H\alpha}=2 and with Silva et al. (2018) model, the Hα\rm H\alpha signals are reconstructed as accurately as the default cHα=1c_{\rm H\alpha}=1 case.

We also compare the one-point PDFs and the power spectra in Figure 5, where we plot the averages over 1000 test data with cHα=0.5c_{\rm H\alpha}=0.5 (cyan line), 1 (red), and 2 (purple), respectively. The yellow lines show those with the Silva et al. (2018) model. The statistics of each reconstructed image are computed by taking the means of those reconstructed by 5 different networks. Both the one-point PDFs and the power spectra are reproduced well except the case of cHα=0.5c_{\rm H\alpha}=0.5 that has effectively small signals. We thus conclude that our network is applicable to maps with different intensity models as long as we have a good understanding of the observational noise and if the line intensities are not too weak compared to the training data.

4 Conclusion

We have developed cGANs that effectively reduce observational noise in line intensity maps. We train the cGANs by using a large set of mock observations assuming a realistic noise level expected for the SPHEREx mission. Our cGANs can reconstruct the point-source positions and the PDF of the intensity maps. The power spectrum is also reconstructed remarkably well, but the accuracy depends on the area/volume assumed for the training data. We have also found that our method is able to reconstruct the signals even if the underlying line intensity model is different from the original training data.

If we combine with another set of networks that efficiently separates signals from different redshifts (Moriwaki et al., 2020), the cGANs developed in this study can extract the emission-line signal from galaxies at an arbitrarily specified redshift from noisy maps. Therefore, using data from multi-frequency, wide-field intensity mapping observations, we can reconstruct the three-dimensional distribution of emission-line galaxies. The intensity peaks detected by our cGANs correspond to bright galaxies and galaxy groups with high confidence, which will be promising targets for follow-up observations. Finally, the reconstructed line intensity map essentially traces the distribution of galaxies and hence of underlying matter, and thus it is well suited for cross-correlation analysis with other tracers. Accurate reconstruction of the statistics such as the one-point PDF and power spectrum as shown in this Letter will allow us to perform cosmological parameter inference and to study galaxy formation and evolution using data from future LIM observations.

We thank the anonymous referee for helpful comments and suggestions. We thank Yasuhiro Imoto for helpful comments. KM is supported by JSPS KAKENHI Grant Number JP19J21379, by Advanced Leading Graduate Course for Photon Science (ALPS) of the University of Tokyo, and by JSR Fellowship, the University of Tokyo. MS is supported by JSPS Overseas Research Fellowships. NY acknowledges financial support from JST CREST JPMHCR1414 and JST AIP Acceleration Research JP20317829.

References

  • Breysse et al. (2017) Breysse, P. C., Kovetz, E. D., Behroozi, P. S., Dai, L., & Kamionkowski, M. 2017, MNRAS, 467, 2996, doi: 10.1093/mnras/stx203
  • Cheng et al. (2016) Cheng, Y.-T., Chang, T.-C., Bock, J., Bradford, C. M., & Cooray, A. 2016, ApJ, 832, 165, doi: 10.3847/0004-637X/832/2/165
  • Cheng et al. (2020) Cheng, Y.-T., Chang, T.-C., & Bock, J. J. 2020, ApJ, 901, 142, doi: 10.3847/1538-4357/abb023
  • Cooray et al. (2019) Cooray, A., Chang, T.-C., Unwin, S., et al. 2019, arXiv e-prints, arXiv:1903.03144. https://arxiv.org/abs/1903.03144
  • Crites et al. (2014) Crites, A. T., Bock, J. J., Bradford, C. M., et al. 2014, Proc. SPIE, 9153, 91531W, doi: 10.1117/12.2057207
  • Doré et al. (2014) Doré, O., Bock, J., Ashby, M., et al. 2014, arXiv e-prints, arXiv:1412.4872. https://arxiv.org/abs/1412.4872
  • Ferland et al. (2013) Ferland, G. J., Porter, R. L., van Hoof, P. A. M., et al. 2013, Rev. Mexicana Astron. Astrofis., 49, 137. https://arxiv.org/abs/1302.4485
  • Ferland et al. (2017) Ferland, G. J., Chatzikos, M., Guzmán, F., et al. 2017, Rev. Mexicana Astron. Astrofis., 53, 385. https://arxiv.org/abs/1705.10877
  • Goodfellow et al. (2016) Goodfellow, I., Bengio, Y., & Courville, A. 2016, Deep Learning (MIT Press)
  • Guo et al. (2013) Guo, Q., White, S., Angulo, R. E., et al. 2013, MNRAS, 428, 1351, doi: 10.1093/mnras/sts115
  • Isola et al. (2016) Isola, P., Zhu, J., Zhou, T., & Efros, A. A. 2016, CoRR, abs/1611.07004. https://arxiv.org/abs/1611.07004
  • Koopmans et al. (2015) Koopmans, L., Pritchard, J., Mellema, G., et al. 2015, in Advancing Astrophysics with the Square Kilometre Array (AASKA14), 1. https://arxiv.org/abs/1505.07568
  • Monaco et al. (2013) Monaco, P., Sefusatti, E., Borgani, S., et al. 2013, MNRAS, 433, 2389, doi: 10.1093/mnras/stt907
  • Moriwaki et al. (2020) Moriwaki, K., Filippova, N., Shirasaki, M., & Yoshida, N. 2020, MNRAS, 496, L54, doi: 10.1093/mnrasl/slaa088
  • Nelson et al. (2019) Nelson, D., Springel, V., Pillepich, A., et al. 2019, Computational Astrophysics and Cosmology, 6, 2, doi: 10.1186/s40668-019-0028-x
  • Planck Collaboration VI (2018) Planck Collaboration VI. 2018, arXiv e-prints, arXiv:1807.06209. https://arxiv.org/abs/1807.06209
  • Shirasaki et al. (2019) Shirasaki, M., Yoshida, N., & Ikeda, S. 2019, Phys. Rev. D, 100, 043527, doi: 10.1103/PhysRevD.100.043527
  • Silva et al. (2018) Silva, B. M., Zaroubi, S., Kooistra, R., & Cooray, A. 2018, MNRAS, 475, 1587, doi: 10.1093/mnras/stx3265
  • Sobral et al. (2013) Sobral, D., Smail, I., Best, P. N., et al. 2013, MNRAS, 428, 1128, doi: 10.1093/mnras/sts096
  • Sobral et al. (2015) Sobral, D., Matthee, J., Best, P. N., et al. 2015, MNRAS, 451, 2303, doi: 10.1093/mnras/stv1076
  • Visbal & Loeb (2010) Visbal, E., & Loeb, A. 2010, J. Cosmology Astropart. Phys, 11, 016, doi: 10.1088/1475-7516/2010/11/016
  • Zhang et al. (2019) Zhang, X., Wang, Y., Zhang, W., et al. 2019, arXiv e-prints, arXiv:1902.05965. https://arxiv.org/abs/1902.05965