Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences
Open AccessResearch articles

Modes of the Dark Ages 21 cm field accessible to a lunar radio interferometer

Philip Bull

Philip Bull

Jodrell Bank Centre for Astrophysics, University of Manchester, Manchester M13 9PL, UK

Department of Physics and Astronomy, University of Western Cape, Cape Town 7535, South Africa

[email protected]

Contribution: Conceptualization, Formal analysis, Investigation, Project administration, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

Google Scholar

Find this author on PubMed

,
Caroline Guandalin

Caroline Guandalin

School of Physical and Chemical Sciences, Queen Mary University of London, London E1 4NS, UK

Contribution: Investigation, Software, Visualization, Writing – original draft

Google Scholar

Find this author on PubMed

and
Chris Addis

Chris Addis

School of Physical and Chemical Sciences, Queen Mary University of London, London E1 4NS, UK

Contribution: Investigation, Visualization

Google Scholar

Find this author on PubMed

    Abstract

    At redshifts beyond z30, the 21 cm line from neutral hydrogen is expected to be essentially the only viable probe of the three-dimensional matter distribution. The lunar far-side is an extremely appealing site for future radio arrays that target this signal, as it is protected from terrestrial radio frequency interference, and has no ionosphere to attenuate and absorb radio emission at low frequencies (tens of MHz and below). We forecast the sensitivity of low-frequency lunar radio arrays to the bispectrum of the 21 cm brightness temperature field, which can in turn be used to probe primordial non-Gaussianity generated by particular early universe models. We account for the loss of particular regions of Fourier space due to instrumental limitations and systematic effects, and predict the sensitivity of different representative array designs to local-type non-Gaussianity in the bispectrum, parametrized by fNL. Under the most optimistic assumption of sample variance-limited observations, we find that σ(fNL)0.01 could be achieved for several broad redshift bins at z30 if foregrounds can be removed effectively. These values degrade to between σ(fNL)0.03 and 0.7 for z=30 to z=170, respectively, when a large foreground wedge region is excluded.

    This article is part of a discussion meeting issue ‘Astronomy from the Moon: the next decades (part 2)’.

    1. Introduction

    The cosmic Dark Ages, roughly corresponding to the redshift range z301000 [1,2], constitute one of the last great frontiers of observational astronomy. This period is sandwiched between the recombination era, when the baryonic content of the Universe became electrically neutral for the first time (z1090), and the Cosmic Dawn (z1530), when the first stars and galaxies formed and began to reionize the neutral intergalactic medium (IGM).

    The large lookback time and lack of luminous sources from this period make any form of direct observation of the Dark Ages challenging. The baryonic matter at this time comes predominantly in the form of neutral Hydrogen (HI) however, which has a distinctive spectral line deep in the radio part of the spectrum, at a wavelength of 21.1 cm (1420.4 MHz) [3]. By the time of observation today, this has been strongly redshifted into the low-frequency end of the radio spectrum, at tens of MHz and below. This is quite fortunate, as the late Universe is optically thin to these frequencies. Radio emission from the Dark Ages can travel, effectively unimpeded, from its time of emission until the present-day, making it a highly promising probe of this otherwise ‘dark’ epoch [3].

    Immediately following recombination at z1090, fluctuations in the baryonic matter density, including the neutral hydrogen gas, had a different amplitude and scale dependence compared with the dominant cold dark matter (CDM) component [4]. Over time, the baryon distribution evolved to match the CDM distribution on sub-horizon scales, down to around the Jeans scale (kJ300Mpc1), where baryon pressure effects take over [4,5]. By measuring the statistical clustering properties of the baryons, we can therefore infer how the dark matter is clustered, which in turn can be related back to the cosmic initial conditions set during the inflationary epoch.

    The three-dimensional spatial distribution of the neutral hydrogen is not observed directly however. The actual observable quantity is the intensity, or equivalently the brightness temperature, Tb, of the HI as a function of angle on the sky and radio frequency. The frequency can be mapped to an observed redshift, which allows us to reconstruct a three-dimensional map of the HI as it is projected onto our past lightcone, i.e. the frequency dimension captures the evolution of the brightness temperature field in both time and comoving distance from us. The peculiar motions of the gas also contribute a Doppler shift, further distorting the distribution in the frequency direction [3]. These effects are expected to be relatively mild, and predictable.

    Dark matter clustering can also be studied using large-scale structure probes at later times. The tracer populations used there are expected to have a complex relationship with the dark matter distribution that they are embedded in however—the connection between galaxies and the dark matter field is difficult to model, even in light of modern hydrodynamical simulations, and so contributes substantial theoretical uncertainties into the interpretation of the observed clustering [6]. This is particularly the case on smaller scales, of order Mpc and below, where dark matter has assembled into collapsed halo objects that are populated by different types of galaxies according to complex non-linear galaxy formation and feedback processes [7]. The effects of nonlinear gravitational collapse influence increasingly large scales as time goes on, breaking the connection between the observed clustering of matter and the primordial fluctuations that initially seeded it.

    The picture is somewhat simpler in the Dark Ages. Nonlinear collapse is confined to much smaller scales—large collapsed objects have not yet had time to form—and there are not yet any galaxies to participate in complicated formation and evolution processes. Nor do we need to worry about complicated and uncertain radiative processes that affect the ionisation properties of the gas during Cosmic Dawn and the subsequent Epoch of Reionization (EoR) between z630. The 21 cm brightness temperature is not simply a linear tracer of the baryon density or CDM density however. The local brightness temperature fluctuation δTb depends on the thermal state of the gas (via the spin temperature, Ts) as well as on the local HI density, and is also modulated by an optical depth term that depends on the line-of-sight peculiar velocity [4,5]. While these terms and their relation to the underlying baryon and CDM fluctuations, δb and δc, can be calculated analytically, the mapping between them necessarily includes terms beyond linear order.

    For cosmological interpretation, the quantities of interest are not the three-dimensional matter field itself, but its statistical properties, as these can be predicted theoretically, regardless of the particular statistical realization of the field that we observe. Existing observational constraints, e.g. from cosmic microwave background (CMB) experiments, point towards a cosmic matter distribution that is approximately statistically homogeneous, isotropic, and close to Gaussian-distributed [8]. The implication of these properties is that, to a very good approximation, the statistics of the matter distribution are fully described by the power spectrum of the matter density fluctuations, i.e. their 2-point function in Fourier space, defined by

    δm(k)δm(k)=(2π)3δ(3)(kk)Pm(|k|).1.1
    Here, the Fourier-space matter density fluctuation is defined through ρm(k)=ρ¯m(1+δm(k)), where ρ¯m is the mean matter density at a given redshift. The angle brackets denote an ensemble average (which can be replaced by a spatial average over sufficiently large volumes), δ(3) is the three-dimensional Dirac delta function, and Pm(|k|) is the total matter power spectrum, which encodes the variance of the Fourier-space baryon + CDM field as a function of wavenumber (inverse distance scale), k|k|. Under the assumption of Gaussianity, all even higher-order moments of the matter distribution can be calculated from products of the power spectrum, while all odd moments vanish. Exact Gaussianity is broken by nonlinear gravitational processes however, which induce higher-order moments and couple different Fourier modes. Importantly, mild non-Gaussianities can also be caused by primordial processes, such as the differential evolution of the fields in multi-field inflation models [9]. The detection of primordial non-Gaussianity (PNG) in the cosmic matter distribution would therefore give us a uniquely valuable way of probing at least some of the physical processes at work in the very earliest moments of the Universe’s history.

    A useful probe of non-Gaussianity is the bispectrum, i.e. the 3-point function of the matter density field in Fourier space, which vanishes for a perfectly Gaussian field. Different types of primordial model predict different amplitudes and shapes of the bispectrum [10]. The most commonly studied is local-type non-Gaussianity, which gives rise to a bispectrum that is largest for ‘squeezed’ 3-point configurations (i.e. where the triangle formed by the three k vectors of the 3-point function are elongated isosceles triangles, k1k2k3). The amplitude of this bispectrum shape is typically parametrized by fNL, and current constraints from the CMB place an upper limit of approximately |fNL|5 [11]. Depending on the inflation model, one can obtain values of fNL1 (for particular multi-field models for instance), down to a strong prediction of fNL103 for single field models [12,13]. Other bispectrum shapes can also be obtained and their amplitudes constrained [14,15].

    The primary CMB temperature fluctuations have been measured well enough that no further (substantial) improvement in bispectrum constraints can be made from this source—the constraints are now dominated by ‘cosmic variance’, which is the intrinsic uncertainty due to having only a finite number of samples of a given quantity (in this case, a finite number of observable Fourier modes in the observed CMB). Late-time large-scale structure probes measure a larger number of Fourier modes than the CMB, and so can be used to further improve constraints and push closer to the σ(fNL)1 level required to test some types of inflationary model, albeit with the difficulty of needing to account for nonlinearity and astrophysical modelling uncertainty [12,1619].

    Pushing to the fNL103 level requires vastly more Fourier modes, however [20]. This is where Dark Ages 21 cm mapping experiments have a crucial role to play. There are very many more modes available from extending to higher wavenumbers than are reasonably accessible to low-redshift surveys, as the number of Fourier modes in a survey volume scales like kmax3, where kmax is the maximum recoverable Fourier wavenumber. Furthermore, the non-primordial contributions to the Dark Ages 21 cm bispectrum, such as those due to the non-linear mapping between the 21 cm signal and the underlying CDM field, can be calculated analytically, with comparatively fewer theoretical uncertainties in how to model them.

    In this article, we examine some of the practical challenges associated with observing the bispectrum of 21 cm brightness temperature fluctuations from the cosmic Dark Ages. The highest redshifts (and a large fraction of the Fourier modes) can only be accessed by radio telescopes outside the Earth’s atmosphere, due to the scattering effect of the ionosphere on low-frequency radio waves, while radio frequency interference from around the Earth is also challenging to identify and remove at the relevant frequencies. Hence, there have been a number of proposals to build a Dark Ages 21 cm instrument on the far side of the Moon, or in lunar orbit.

    We use the technique of Fisher forecasting to obtain simplified predictions for how well different array configurations should be able to measure the HI bispectrum. We incorporate the effect of losing modes to systematic effects such as radio foreground emission in our forecasts, as well as the intrinsic limitations due to instrumental resolution.

    2. Lunar radio array configurations

    Leaving matters such as deployment, power, data transmission and array calibration aside, the main properties that determine the observing characteristics of a radio interferometer array are its frequency range (bandwidth) and spectral resolution, the field of view of individual receiving elements (also called the primary beam), and the distribution of baselines, i.e. the number of available correlated antenna pairs as a function of the length and orientation of the separation vector between them. All of these terms are represented in the noise power spectrum, PN (see equation (3.11) below). In this section, we attempt to distill the various lunar arrays that have been proposed into a handful of model configurations that we can then use to produce representative forecasts.

    (a) Fourier-space sampling function

    First, we briefly review the connection between the instrumental configuration, including the baseline distribution, and the set of Fourier modes on the sky that can be observed by the interferometer and thus included in the bispectrum measurements. An illustration for some representative instrumental properties is shown in figure 1.

    Figure 1.

    Figure 1. ‘Exclusion’ plot showing which cylindrical Fourier modes (k,k) are observable with an interferometer for four representative redshifts. Shaded regions would be excluded from the observations due to various observational effects: the foreground wedge (grey, diagonal); the frequency channel width (grey-blue, top); the fractional bandwidth of the observation (magenta, bottom); and the minimum and maximum baseline lengths (blue, left and yellow, right, respectively). We have assumed representative numbers here: 10 kHz frequency channels, a 30% effective bandwidth at each redshift, and minimum and maximum baselines of 10 m and 10 km, respectively.

    (i) Baseline distribution

    The baseline distribution can be represented as a number density of baselines in the uv-plane. Vectors in this plane can be mapped into transverse Fourier vectors by k=2πu/χ(z) (where χ is the comoving distance to redshift z), assuming that k has been defined in a plane on the sky that is parallel to the plane of the interferometer array, which we take to be perfectly flat, tangent to the lunar surface, with a normal that points at the zenith. Under these assumptions, the uv-plane baseline distribution can be calculated by finding the vector u=d/λ corresponding to each baseline, where d is the separation vector between the antennae. The set of baseline vectors can then be binned appropriately in the uv-plane to obtain the number density nb. Note that the uv-plane distribution changes with observing wavelength since u1/λ, and so nb(u) will also vary with redshift.

    In the flat-sky limit, each baseline can be thought of as being sensitive to the amplitude (in brightness temperature) of a single Fourier mode on the sky with wavevector k. The field of view of the antennae acts as a window function on the set of Fourier modes that the array is sensitive to however, and so introduces a correlation scale in the uv-plane of approximately Δu1/ΩFOV. This scale can be used as the minimum bin width for the baseline number density distribution, and also sets a lower limit on the minimum recoverable k mode (i.e. maximum recoverable angular scale).

    Importantly, only Fourier modes on the sky represented by baselines that are present in the array can be recovered. Unrepresented Fourier modes are not measured, and therefore cannot be used in the bispectrum measurements, etc. In Earth-based interferometry applications, it is common to use the rotation of the baseline vectors with respect to the sky, due to the Earth’s diurnal rotation, to perform rotation synthesis. As the Earth rotates, each baseline migrates along an elliptical track in the uv-plane as its orientation and projection of the baseline changes with respect to a reference point on the sky. This allows a wider range of Fourier modes to be recovered by each baseline as the rotation progresses.

    For lunar applications, the situation is different, as there is no diurnal rotation. The Moon orbits the Earth during the lunar month, and the Earth orbits the Sun, so observations taken across a terrestrial year will allow some degree of rotation synthesis to be achieved [21]. Different patches of the sky will also rise and set throughout the month. Since the Sun is a bright radio source, observations would normally be taken during the lunar night however, and so this prevents some segments of the tracks in the uv-plane from being recovered.

    (ii) Foreground wedge

    Radio interferometer visibilities measure the integrated sky intensity distribution after it has been modulated by a baseline-dependent fringe pattern. To a good approximation, the fringe pattern for each baseline can be mapped to a transverse Fourier mode, k, at each frequency. The sky intensity distribution is also modulated by the primary beam pattern of the antennae however, which also depend on frequency, but in a different way. When a Fourier transform is performed in the frequency direction, i.e. to give the visibility in terms of radial Fourier mode, k, this additional modulation gives rise to a coupling between transverse and radial modes of the sky signal. This scatters intrinsically spectrally-smooth sky emission (i.e. the foregrounds) at low-k into an extended wedge-shaped region in (k,k) space. The maximum extent of the wedge region is in principle related to the maximum geometric delay in arrival time of a wavefront between the two antennae of a baseline, which occurs when a source is on the horizon. This ‘worst case’ foreground-contaminated region is referred to as the ‘horizon wedge’.

    The intensity of the contamination can vary within the wedge, depending on the shape of the primary beam pattern—for dipole antennae the wedge is quite evenly contaminated, for instance, while for parabolic reflectors, it tends to be more localized into the ‘prongs’ of a pitchfork shape, with cleaner regions in between them. Since most lunar array concepts employ dipoles for reasons of cost and simplicity, we can anticipate severe contamination throughout the entire wedge region. Cleaning foreground emission from inside the wedge region requires exceptionally good models of both the primary beams and the foreground emission however, and many ground-based arrays choose a more conservative ‘avoidance’ strategy instead. This is where 21 cm signal modes within the wedge region are assumed to be irretrievable and so are excised, while a variety of analysis choices are made to prevent power from leaking out of the wedge region into an otherwise clean ‘window’ region.

    In the absence of a detailed plan to support extraction of 21 cm signal modes within the wedge region by characterizing the foregrounds and beam patterns of a lunar array with extreme precision, we can conservatively assume that modes within the horizon wedge are unusable for 21 cm cosmology (figure 1). More optimistically, calibration methods such as [22] may allow substantial suppression of the foregrounds however. We bracket the possibilities by producing forecasts for both horizon-wedge (pessimistic) and no-wedge (optimistic) scenarios.

    (iii) Bandwidth and spectral resolution

    The frequency axis of the observations simultaneously encodes the evolution of the signal with redshift, z, and the variations of the 21 cm field in the radial or line-of-sight direction (represented by the radial Fourier wavenumber, k). It is common to make an approximation that the redshift evolution is negligible over a small frequency range (i.e. within a sufficiently narrow redshift bin), so that frequency channels can be mapped to a radial distance within a three-dimensional volume at a ‘fixed’ central redshift. The frequency resolution then gives the maximum radial Fourier wavenumber kmax that can be measured in the volume. Similarly, the bandwidth over which the redshift evolution is neglected can be converted into a maximum radial extent of the three-dimensional volume, and so sets the fundamental radial mode, kmin.

    While there are methods that allow this redshift-binning approximation to be avoided, working with the Fourier-space power spectrum and bispectrum typically requires some kind of redshift binning to be done, and so we take it as an unavoidable aspect of our analysis here. The redshift bin width can be chosen to keep the cosmological evolution of the signal across the bin sub-dominant, and so will typically vary with frequency. As a somewhat maximal choice, we assume a 30% fractional bandwidth within each bin. For bins centred at z=10,40,100,200 (ν130,35,14,7 MHz), this would equate to bandwidths of Δν40,10,4,2 MHz.

    (b) Representative baseline distributions

    In this section, we attempt to distill the various mission concepts in the literature into a set of representative development stages for a lunar 21 cm array. We have used the CoDEX [23], FarSide/FarView [24,25] and ROLSS/DALI [26] concepts as the basis for the following.

    (i) Stage I

    A relatively small pathfinder array with a dense core, likely in a grid layout that is the minimum needed to test antenna deployment technologies and implement an FFT-based correlator, e.g. a 16×16 (=256 antennae) or 32×16 (=512 antennae) array. To soften the technical requirements, choices such as a reduced bandwidth (e.g. 20–60 MHz) and relatively short maximum baseline length can be made. For maximum and minimum baseline lengths of approximately 1 km and 30 m, respectively, scales in the range 5×103k5×102Mpc1 would be accessible, i.e. around the top end of the BAO scale.

    (ii) Stage II

    A large array of around 10 000 antennae with a large core plus some outriggers to increase the maximum baseline length to around 5 km. Building on the technology demonstrated by the Stage I instrument, this could allow for deviations from an FFT-friendly regular grid into a more balanced layout, permitting longer baselines for imaging. Improvements in data rate, antenna design etc. would allow a wider bandwidth to be observed, perhaps in the 10–100 MHz range, covering both Cosmic Dawn and well into the Dark Ages, 13z140 (note that we only consider z20 in this paper). Angular scales in the range 5×104k5×101Mpc1 would be accessible with this instrument, permitting some initial imaging applications at the lowest redshifts.

    (iii) Stage III

    A giant array of 100 000 or more antennae over a large area of 10×10km or so. This would further improve on the frequency range of Stage II, probing frequencies down to 5 MHz or so, and angular scales in the range 104k100Mpc1.

    In all cases, we employ a minimum baseline length of 20 m. This should be compared with the maximum wavelength of between approximately 1560 m (at 20 MHz and 5 MHz, respectively), and a typical dipole length of order approximately 10 m. Neighbouring antennas will be within the near-field of one another at higher frequencies, and we can expect mutual coupling effects to be important.

    Rather than defining an explicit baseline distribution for each stage, we attempt to capture representative distributions based on a fitting function for the circularly averaged density,

    n(d)=A(DD0)2exp[(DD0w)2],2.1
    where A is a normalizing constant determined by the total number of unique baselines, DminDmax2πDn(D)dD=Nant(Nant1)/2, where the number of antennae is given in table 1. The function was designed to give a good fit to the FarView concept’s baseline distribution, which has a regular grid in the core and semi-random outlier stations [25]. For Stages I, II and III, we chose the parameters D0=180,1000,2100 m and w=310,3100,6200 m, respectively. These values give maximum baseline lengths around the right value for the specifications in table 1. The resulting baseline distributions are shown in figure 2.
    Figure 2.

    Figure 2. Notional baseline distributions for the three representative arrays. The functional form is chosen to have a similar form to the one for FarView for the Stage III experiment. Vertical dashed lines show the maximum baseline length for a regular square array with linear extent 500 m, 5 km and 10 km, respectively. The Stage I and II experiments have similar densities for the shortest baselines, differing only in the number of long baselines, consistent with extending a dense core by adding outriggers at intermediate and large distances.

    Table 1. Representative specifications of the three development stages for lunar 21 cm arrays. These numbers are loosely aligned with mission concepts in the literature, including ones that propose a staged approach.

    configuration no. antennas approx. area freq. range redshift range
    Stage I O(500) 1 km × 1 km 20–60 MHz 23–70
    Stage II O(104) 5 km × 5 km 10–60 MHz 23–140
    Stage III O(105) 10 km × 10 km 5–60 MHz 23–280

    3. Fisher forecasting formalism

    In this section, we describe the theoretical calculation of the 21 cm power spectrum and bispectrum during the dark ages, and the Fisher forecasting formalism that we use. We base our calculations on [5,27,35].

    The Fisher matrix for the anisotropic bispectrum can be calculated as

    Fαβ(B)(zi)=14πtriangles1+1dμ102πdϕ1ΔBHI2(zi)BHI(zi)pαBHI(zi)pβ,3.1
    where BHI(zi)BHI(k1,k2,k3,zi) is the HI brightness temperature in redshift bin i with central redshift zi, and the cosine μ1 and angle ϕ parametrize different orientations of the triangle with respect to the line of sight. Under the assumption of Gaussianity of the covariance, the variance of the bispectrum for a given triangle configuration is given by
    ΔBHI2(k1,k2,k3,zi)=s123πk1k2k3(kfΔk)3PHI(k1,μ1,zi)PHI(k2,μ2,zi)PHI(k3,μ3,zi),3.2
    where s123 is a multiplicity factor (equal to 6 for equilateral triangle configurations, 2 for isosceles, and 1 for scalene), and kf2π/L is the fundamental mode in Fourier space for an (assumed cubic) survey volume of comoving side length L. We will take the continuum limit of equation (3.1) and extend the integrals over the full range of k1,k2 and k3. For that, we divide equation (3.1) by the total number of equivalent triangles r123. Since r123=(1,3,6) for equilateral, isosceles and scalene configurations, respectively, r123s123=6. Hence, we consider the following expression for the Fisher matrix:
    FαβB=V4π216dμ1dϕkminkmaxdk1dk2dk3k1k2k3(2π)3×αBHI(zi)βBHI(zi)PHI(k1,μ1,zi)PHI(k2,μ2,zi)PHI(k3,μ3,zi).3.3
    We stress that the continuum approximation may overestimate the total number of triangles used to forecast the errors on fNL; in reality, one obtains fewer triangles per configuration due to the binning strategy in Fourier space. For example, one can access fewer triangles with thicker bins, which reduces the expected bispectrum signal [28]. Therefore, the results presented in §4 should be taken as a lower bound for the errors expected with the experiments considered here. We also stress that we are ignoring nonlinear corrections to the variance, ΔBHI2.

    (a) Model of the HI bispectrum during the Dark Ages

    Our model for the HI bispectrum is given by (following [5])

    BHI(k1,k2,k3)=BHIprim(k1,k2,k3)+BHIgrav(k1,k2,k3)+BHInl(k1,k2,k3),3.4
    where BHIprim is related to the primordial bispectrum of scalar perturbations, BΦ, as
    BHIprim(k1,k2,k3)=F(k1,k2,k3,μ1,μ2,μ3)BΦ(k1,k2,k3),3.5
    BHIgrav is the gravitational contribution to the bispectrum,
    BHIgrav(k1,k2,k3)=G(k1,k2,μ1,μ2)PmL(k1)PmL(k2)+2 perms.,3.6
    and BHInl accounts for the nonlinear relation between the brightness temperature and the baryon density
    BHInl(k1,k2,k3)=H(k1,k2,μ1,μ2)PmL(k1)PmL(k2)+ 2 perms.3.7

    The full expressions for these terms are given in [5], but we briefly describe them here. The functions F,G and H depend on the angle between the wavevectors ki and the line of sight n^, μi=kin^/ki, on the mean 21 cm brightness temperature at redshift z, T¯HI, and on the derivative of the brightness temperature THI with respect to the linear baryon overdensity δb(1): αTHI/δb(1). For the primordial bispectrum, F also carries a dependence on M(k,z)2k2T(k)D(z)/3ΩmH02 (e.g. see [10]); for the gravitational contribution, G also depends on the second-order perturbation theory kernels, F2 and G2, and on γTHI/δb(2); and for the nonlinear part, H also depends on β122THI/δb2. Finally, the primordial bispectrum BΦ can be constructed from a sum of contributions from different bispectrum shapes; we retain only the local-type bispectrum in this work, which can be calculated as

    BΦlocal(k1,k2,k3)=2fNL[PΦ(k1)PΦ(k2)+PΦ(k1)PΦ(k3)+PΦ(k2)PΦ(k3)].3.8
    Here, PΦ is the primordial scalar power spectrum, and fNL is the overall amplitude of the local-type bispectrum, as described above. Note that [5] uses the flat-sky approximation and neglects the Wouthuysen–Field effect. An example of some of the different contributions to the bispectrum at z=30, including the uncertainty for a cosmic variance-limited scenario, is shown in figure 3.
    Figure 3.

    Figure 3. Amplitude of the bispectrum at z=30, for triangle configurations where one of the sides is fixed to k1=0.05Mpc1. (a) The primordial bispectrum as a function of triangle configuration, assuming local-type non-Gaussianity with fNL=1 (note the higher amplitude in the top left). (b) The gravitational contribution to the bispectrum, which is significantly larger. (c) The uncertainty on the bispectrum assuming no thermal noise or instrumental effects (cosmic variance contribution only).

    In a more comprehensive treatment, we could calculate the Fisher matrix for a range of cosmological parameters in order to account for their uncertainties, and correlations between the different parameters. This is largely a matter of evaluating the derivatives of equation (3.4) with respect to these parameters, as they appear in equation (3.1). In the present article, however, we focus only on the simplest case of an idealized forecast for a single parameter, fNL. Since this parameter appears as a prefactor of only a single term in equation (3.4), the expression for the derivatives simplifies to

    BHIfNL=BHIprimfNL=FBΦlocalfNL.3.9
    Implicit in this simplified treatment is an assumption that we are able to perfectly subtract the gravitational component of the bispectrum, which is typically 12 orders of magnitude larger than the primordial component (assuming fNL1).

    The final ingredient needed to evaluate the Fisher matrix for fNL is an expression for the dark ages 21 cm power spectrum, which appears in equation (3.2). This is given by

    PHI(k,μ,z)=(α+T¯HIμ2)2Pδb(k)+PN(k,z),3.10
    where Pδb is the linear power spectrum of baryon fluctuations.

    (b) Model of the interferometer noise power spectrum

    The last factor, PN, is the power spectrum of the instrumental noise. This term describes the noise variance on each Fourier mode measured by the radio array. This sets the noise level on each triangle configuration that contributes to the bispectrum.

    Following the notation of [27], the noise power spectrum for an interferometer can be calculated from the radiometer equation as

    PN(k,z)=Tsys2λ(1+z)χ2(z)H(z)(λ2Aeff)21Npolnb(u)ttotSareaΩFOV,3.11
    where Tsys=Tsky+TinstTsky is the system temperature, H(z) and χ(z) are the expansion rate and comoving radial distance to redshift z, λ is the observed wavelength of the redshifted 21 cm (i.e. λ=0.211(1+z)m), Aeff is the effective area of each receiving element of the radio interferometer, ΩFOV is the solid angle of the beam pattern of each element, Npol=1 or 2 is the number of available receiver polarization channels, and ttot and Sarea are the total observing time for the survey and the total survey area, respectively. We model the sky temperature as Tsky5000K(ν/νref)2.5, where νref=50MHz. Finally, nb(u) is the number density of interferometer baselines in a region of the uv-plane, where u is a two-dimensional vector that can be mapped to a transverse Fourier mode on the sky, k.

    Figure 4 shows the noise power spectra of the three stages of experiment, using the idealized baseline distributions discussed in §2. As a reference survey timescale, we assume ttot=22000 h of observing time, which crudely represents an efficient 5-year survey with a 50% duty cycle, e.g. to account for flagging of data when the Sun is up or bright sources are in the sidelobes. The survey area is assumed to be 20000deg2 in each case, with a field of view of ΩFOVπ steradians (10300deg2). For a dipole-like antenna, the effective area is Aeff=λ2G/(4π), where we assume a slightly enhanced gain of G2. The sharp cutoff at low k is due to the assumed minimum baseline length of 20 m, although lower k could potentially be recovered if calibrated zero-spacing (autocorrelation) data can be obtained. There is a clear redshift dependence, with a shift to lower k as z increases due to the frequency dependence of the fringe pattern of each baseline, and a shift to higher noise power as z increases (frequency decreases) due to the increase in system (sky) temperature. As the array increases in size, the number of baselines increases on all scales, increasing the sensitivity by an order of magnitude or more between each generation. Larger arrays also have longer maximum baseline lengths, and so reach a higher effective maximum k.

    Figure 4.

    Figure 4. Noise power spectrum PN(k,z), as given in equation (3.11), assuming ttot=22000 h of observation time (equivalent to 5 years with a 50% duty cycle). The curves correspond to the three types of experiment: Stage I (blue), Stage II (red) and Stage III (orange), with solid lines denoting redshift z=30 and dotted lines z=60. The increase in array size gains between one and two orders of magnitude in sensitivity between each stage. For comparison, the model considered in [29] has a dimensionless power spectrum of Δ2=0.44mK2 at k=0.1Mpc1 and z=51 (about 0.025K2h3Mpc3, i.e. below the detection threshold even for Stage III with these figures).

    In what follows, we will assume that the noise power spectrum is sub-dominant, i.e. that we are in the sample variance-dominated limit. Reaching this limit would be a major technical feat, involving a very long survey duration, excellent control over systematic effects, calibration errors, and so on. It is nevertheless useful to consider this limit as giving the best constraints that could possibly be achieved with a given array configuration.

    4. Results

    In this section, we present Fisher matrix forecasts for the fNL parameter measured from the HI brightness temperature bispectrum during the Dark Ages, using the formalism described in §3. We consider the three stages of experiment described in table 1, and show the impact of different assumptions about instrumental/scale cuts related to foreground contamination etc.

    For all of the forecasts, we assume a flat ΛCDM model with parameters h=0.674,ns=0.965,σ8=0.811,Ωm=0.315 and Ωb=0.049 [30]. We restrict ourselves to the redshift range z=23200 (frequencies between 7 and 60 MHz) in order to keep the same fitting functions for the HI brightness temperature etc. given in [5], and do not marginalize over uncertainties in astrophysical quantities that set the overall scale of the HI signal, or cosmological parameters that set the shape of the non-primordial contributions to the bispectrum. These assumptions will be relaxed in future work, and have been studied in past works, e.g. [5,29]. As explained in §2, we also make assumptions about the co-planarity of the baselines, and use a flat-sky limit that also neglects cosmological evolution within discrete redshift bins. In all the cases presented here, the redshift bins are chosen to be sub-bands with 30% bandwidth at their respective centre frequencies. This is at the upper end of what is plausible if we wish to neglect cosmic evolution within each bin.

    Figure 5 shows the forecast 68% CL error on the fNL parameter for each generation of experiment as a function of redshift. Under the assumptions, we have made—particularly neglecting the thermal noise contribution—all three configurations produce similar results in redshift bins where they overlap.1 The main difference between the three stages largely comes down to the maximum redshift that each of them can reach. In practice, the Stage I array has about 300 times the noise level of the Stage III array at the same k, and so would need to integrate for a much longer time to reach the same signal to noise ratio.

    Figure 5.

    Figure 5. Cosmic-variance limited (PN=0) Fisher forecast for the fNL parameter for local-type primordial non-Gaussianities using the three stages of lunar 21 cm experiments. The results are shown for the foreground-free case (solid), i.e. kmin=2π/Vzi1/3, and with foregrounds (dashed), i.e. excluding the horizon wedge from the data. The channel width (δν) sets the maximum radial scale kmax included in the analysis. (Note that we use this as the maximum radial scale rather than a nonlinear cut-off, as would be more usual for low-z experiments.)

    A number of effects contribute to the shape of the curves in figure 5. The comoving volume of each redshift bin is a key factor in setting the sample variance limit, and increases significantly from low to high redshift. This is tempered by the increasing Tsys with redshift. Recall that fNL encodes the amplitude of the primordial bispectrum in the squeezed limit, where there is one low-k leg and two higher-k legs to each triangle. The number of large-scale modes available to form the low-k leg of the squeezed limit triangles also increases with z. Redshift-dependent limits on σ(fNL) of better than 102 can be obtained at z20, surpassing the CMB and low-redshift galaxy surveys. In the case where the foreground wedge region must be excised completely from the data (figure 1), this degrades to around 0.05 at z50 and 0.5 at z100, which is still competitive with current constraints. Significantly more evolution in σ(fNL) with redshift is observed when the wedge is removed, as the size of the wedge region depends on wavelength.

    We stress that these forecasts are optimistic in a number of ways. Most importantly, the secondary/gravitational bispectrum and astrophysical prefactors (i.e. T¯HI,α,β,γ) have been assumed to be perfectly known. In [5], these terms were found to be important however, with the astrophysical prefactors exhibiting strong correlations with fNL in their Fisher forecasts. Their bottom-line forecast for fNL was weaker than our prediction, with σ(fNL)0.23 predicted for a large array (other differences, such as their choice of larger 100 kHz channel widths, also contribute to the discrepancy). We note, however, that relatively mild priors on the astrophysical factors, e.g. from models fitted at lower redshift, should be helpful in breaking the strong correlations, potentially allowing better constraints on fNL to be achieved than presented in [5]. In this case, our results should be considered to bound the possible values of σ(fNL) from below, as we are operating in the optimistic sample variance-limited case.

    5. Conclusion

    In this paper, we reviewed the distinctive properties of the Dark Ages 21 cm brightness temperature field as a probe of early Universe physics, particularly as encoded by the local-type non-Gaussianity parameter fNL. We examined how a lunar radio interferometer could be used to measure the 21 cm bispectrum while avoiding severe problems such as ionospheric distortion and radio frequency interference on Earth, and discussed how different instrumental effects contribute scale cuts that limit the number of Fourier modes of the three-dimensional 21 cm field that can be recovered. We then went on to define a set of three representative stages of the development of such lunar arrays, beginning with smaller ones with a few hundred antennae spread over a square kilometre, and culminating in a much larger array of a hundred-thousand or more antennae spread over 100km2. These were used in simple exploratory Fisher matrix forecasts to show how well the 21 cm bispectrum, and hence the fNL parameter, could be measured under optimistic assumptions, such as neglecting thermal noise (i.e. obtaining sample variance-limited observations). In essence, our results show the best that a notional lunar 21 cm experiment could do in the absence of systematic effects in the parts of the data that remain after a series of scale cuts, and without limits on observing time.

    We found that severe scale cuts that are applied to many ground-based 21 cm experiments (at lower redshift) in order to remove foreground contamination did degrade the predicted constraints on fNL substantially, but that under our assumptions the signal could still be measured at a level that is competitive with CMB and galaxy survey experiments.

    Overall, we suggest that a staged deployment of lunar 21 cm arrays (moving from compact with a few hundred antennae, to widely distributed with up to a hundred-thousand antennae) should provide a robust path to measuring the local-type non-Gaussianity parameter fNL, with the prospect of substantially improved precision compared to ground-based experiments. While there are obvious engineering and cost challenges associated with deploying such arrays on the Moon, we have seen that the cosmological performance of the arrays survive even quite stringent scale cuts to remove foreground contamination and the like. In terms of future work, it would be desirable to go beyond the analytic approach that we have used here, and perform a direct demonstration of 21 cm bispectrum recovery on simulated data that include realistic instrumental effects such as foregrounds, calibration errors, antenna variations and mutual coupling. We have also commented on the need to marginalize over astrophysical uncertainties, which is likely to further reduce the forecasted precision.

    Data accessibility

    This article has no additional data.

    Declaration of AI use

    We have not used AI-assisted technologies in creating this article.

    Authors' contributions

    P.B.: conceptualization, formal analysis, investigation, project administration, software, supervision, validation, visualization, writing—original draft, writing—review and editing; C.G.: investigation, software, visualization, writing—original draft; C.A.: investigation, visualization.

    All authors gave final approval for publication and agreed to be held accountable for the work performed therein.

    Conflict of interest declaration

    We declare we have no competing interests.

    Funding

    We acknowledge the Royal Society for travel support. This result is part of a project that has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement no. 948764). C.G. is supported by STFC consolidated grant ST/T000341/1.

    Acknowledgements

    We are grateful to T. Flöss, D. Karagiannis, D. Meerburg, G. Orlando and J. Silk for useful discussions, and two anonymous referees for their helpful comments. This work made extensive use of the public code class [31], and the following python packages and libraries: numpy [32], scipy [33] and matplotlib [34].

    Footnotes

    1 In this paper, the following redshift bins are defined such that each comprises around 30% bandwidth at the respective centre frequencies: (1) z=22.732.8; (2) 32.847.3; (3) 47.368.0; (4) 68.097.6; (5) 97.6139.9; (6) 139.9200.2. Only the first 3 and 5 are present for the Stage I and Stage II experiments, respectively.

    One contribution of 13 to a discussion meeting issue ‘Astronomy from the Moon: the next decades (part 2)’.

    Published by the Royal Society under the terms of the Creative Commons Attribution License http://creativecommons.org/licenses/by/4.0/, which permits unrestricted use, provided the original author and source are credited.