Brought to you by:

Articles

A NEW APPROACH TO DETERMINE OPTICALLY THICK H2 COOLING AND ITS EFFECT ON PRIMORDIAL STAR FORMATION

, , , , and

Published 2015 January 21 © 2015. The American Astronomical Society. All rights reserved.
, , Citation Tilman Hartwig et al 2015 ApJ 799 114 DOI 10.1088/0004-637X/799/2/114

0004-637X/799/2/114

ABSTRACT

We present a new method for estimating the H2 cooling rate in the optically thick regime in simulations of primordial star formation. Our new approach is based on the TreeCol algorithm, which projects matter distributions onto a spherical grid to create maps of column densities for each fluid element in the computational domain. We have improved this algorithm by using the relative gas velocities to weight the individual matter contributions with the relative spectral line overlaps, in order to properly account for the Doppler effect. We compare our new method to the widely used Sobolev approximation, which yields an estimate for the column density based on the local velocity gradient and the thermal velocity. This approach generally underestimates the photon escape probability because it neglects the density gradient and the actual shape of the cloud. We present a correction factor for the true line overlap in the Sobolev approximation and a new method based on local quantities, which fits the exact results reasonably well during the collapse of the cloud, with the error in the cooling rates always being less than 10%. Analytical fitting formulae fail at determining the photon escape probability after formation of the first protostar (error of ∼40%) because they are based on the assumption of spherical symmetry and therefore break down once a protostellar accretion disk has formed. Our method yields lower temperatures and hence promotes fragmentation for densities above ∼1010 cm−3 at a distance of ∼200 AU from the first protostar. Since the overall accretion rates are hardly affected by the cooling implementation, we expect Pop III stars to have lower masses in our simulations, compared to the results of previous simulations that used the Sobolev approximation.

Export citation and abstract BibTeX RIS

1. INTRODUCTION

The first stars in the universe (the so-called Population III or Pop III stars) emerged several hundred million years after the Big Bang and dramatically changed the physical conditions of their environment. The properties of Pop III stars have a fundamental influence on many subsequent physical processes, such as the synthesis of heavy elements, subsequent star and galaxy formation, or reionization of the intergalactic gas, and it is therefore crucial to understand primordial star formation. Although a consistent and widely accepted formation scenario developed over the last few years (Glover 2013; Bromm 2013; Greif 2014b), there are still a number of open questions whose answers might modify this picture. A review of some of these open questions is given by Glover et al. (2008). In this paper, we focus on the determination of the optically thick H2 cooling rate, because only little progress has been made on this topic since the late eighties in the context of star formation. Furthermore, H2 emission is the dominant cooling process in this regime and hence a detailed understanding of its efficiency under different physical conditions is key to model Pop III star formation and gas dynamics in primordial halos.

Following the standard formation scenario, the first stars form at redshifts z  ≃  20–30 in dark matter halos that have total masses of M ≃ 106M and virial temperatures of around 1000 K (Bromm & Larson 2004; Glover 2005; Bromm et al. 2009; Clark et al. 2011a). Gas in the center of these dark matter halos decouples and undergoes self-gravitating collapse (Yoshida et al. 2006). The decoupled gas clouds have Jeans masses between MJ = 200 M and MJ = 1000 M (Abel et al. 2000, 2002; Yoshida et al. 2006; Clark et al. 2008; McKee & Tan 2008; Turk et al. 2009; Clark et al. 2011b; Hirano & Yoshida 2013) and a critical number density of n = 104–105 cm−3 when they decouple from their host halos (Abel et al. 2000; Clark et al. 2011a; Bromm 2013). During the collapse, gas cools mainly via H2 rotational and vibrational line emission, with the H2 fractional abundance being a few times 10−3 at the beginning of the collapse and becoming close to unity due to three-body H2 formation above 108 cm−3. Because these cooling processes can cool the gas very efficiently, the overall collapse proceeds almost isothermally until the gas becomes very optically thick at high densities and an adiabatic core forms. The exact determination of the optically thick cooling rate requires information about the velocity profile of the cloud because relative velocities Doppler-shift the spectral lines and therefore increase the photon escape probability. The shortcomings of commonly used methods for optically thick cooling lie mainly in the assumption of isotropy and in the dependence on local quantities.

Due to tidal forces and an initial angular momentum, the collapse does not proceed in a spherically symmetric manner, but rather leads to the formation of a rotationally supported disk around the first protostar. The disk generally extends out to 400–1000 AU, has a characteristic temperature of 1500–2000 K, becomes gravitationally unstable, and fragments into multiple parts (Stacy et al. 2010; Clark et al. 2011b; Hosokawa et al. 2011; Greif et al. 2011, 2012, 2013; Glover 2013; Latif et al. 2013). The disk-like structure is a typical feature due to the inability of the halo to transfer angular momentum outward quickly enough (Smith et al. 2011).

One of the main goals of simulations of primordial star formation is to establish the form of the Pop III initial mass function (IMF). The mass of a Pop III star is the crucial parameter which defines its luminosity, temperature, spectrum, lifetime, final fate, and its metal yields. However, the significant variations and uncertainties in the expected mass ranges (Bromm et al. 2001; Omukai & Palla 2001, 2003; Omukai & Yoshi 2003; Johnson & Bromm 2006; McKee & Tan 2008; Ohkubo et al. 2009; Clark et al. 2011a; Greif et al. 2011; Hosokawa et al. 2011; Hirano et al. 2014) reveal the lack of understanding of the primordial gas' fragmentation behavior. Since cooling is the key ingredient for disk instabilities and fragmentation, an accurate determination of optically thick H2 cooling is of general interest for the determination of the primordial IMF.

Here, we analyze the dependence of fragmentation on two different cooling implementations: the commonly used Sobolev approximation and a more sophisticated, yet computationally expensive approach based on the TreeCol algorithm of Clark et al. (2012). We also present a new approximate method, which yields photon escape fractions with mean relative errors smaller than 10%.

This paper is ordered as follows: In Section 2, we review the different approaches that are commonly used to model optically thick H2 cooling. In Section 3, we describe our numerical methodology. In Section 4, we test the methods and present our results. We discuss these results in Section 6, and conclude in Section 7.

2. COOLING METHODS

In this section, we present the commonly used photon escape probability approach for the determination of optically thick H2 cooling rates and several methods of calculating this probability. Moreover, we derive a correction factor for the Sobolev approximation to account for the overlap of spectral lines.

2.1. Optically Thick Cooling

Since H2 is a symmetric, diatomic molecule, it has no permanent dipole moment and can therefore only radiate via ro-vibrational transitions (Le Bourlot et al. 1999). In the optically thin regime (n ⩽ 109 cm−3) we expect all radiation to escape the cloud freely and the cooling rate can be calculated fairly accurately as a function of the local density, temperature, and chemical composition of the gas. A number of different parameterizations of the H2 cooling rate in terms of these quantities are available in the literature. In this work, we use the rates given in Glover & Abel (2008). In the optically thin case we assume that photons, which are emitted in the ro-vibrational transitions, can escape the cloud without being scattered or absorbed and therefore transport thermal energy outward efficiently. In the optically thick case, however, scattering and absorption events might capture the photons and hence decrease the overall cooling efficiency. One generally assumes complete redistribution of the frequency between these scattering events and uses the "escape probability method" in order to solve this problem. Most important for our further discussion are the applications of the escape probability method by Yoshida et al. (2006) and Clark et al. (2011a) to the case of optically thick cooling in primordial gas. The cooling rate in an optically thick medium is given by

Equation (1)

where Elu = hνlu is the energy separation between the lower level l and upper level u, Alu is the spontaneous radiative transition rate for transitions between u and l, βesc, lu is the probability for an emitted line photon to escape without absorption, and nu is the population density of hydrogen molecules in the upper level. Following Yoshida et al. (2006), we assume all energy levels to be populated according to local thermodynamic equilibrium and we consider rotational levels from J = 0 to 20 and vibrational levels ν = 0, 1, 2. At the temperatures of interest, the contribution of other levels will be negligibly small (<10−3). We take values for the level energies from the compilation made available by P. G. Martin on his Web site1 and the radiative transitions rates from Wolniewicz et al. (1998). Since all other quantities are known for typical conditions in primordial gas, the remaining task is the determination of the photon escape probability.

2.2. Sobolev Approximation

Based on the work by Zanstra (1934), Sobolev (1947) derived the escape probability for constant velocity gradients in his study of expanding envelopes (for an English translation see Sobolev 1960). This approximation has been reviewed during the last few decades (Castor 1970; Hummer & Rybicki 1982; Yoshida et al. 2006) and is widely used in simulations of primordial star formation (Yoshida et al. 2006, 2008; Turk et al. 2011; Clark et al. 2011a, 2011b; Greif et al. 2011; Wolcott-Green et al. 2011; Greif et al. 2012; Hirano & Yoshida 2013; Stacy et al. 2013; Stacy & Bromm 2013; Greif et al. 2013; Hirano et al. 2014; Greif et al. 2014a; Stacy & Bromm 2014). Following the derivation by Emerson (1996), the photon escape probability is given by

Equation (2)

where τlu is the opacity at the line center. The absorption coefficient for a transition from the lower to the upper level is given by

Equation (3)

where Blu is the Einstein coefficient for absorption and ϕ(ν) is the line profile function. The opacity can be written as

Equation (4)

where Lchar is a characteristic length scale (Castor 1970; Goldreich & Kwan 1974; de Jong et al. 1975; Stahler & Palla 2005). According to Equation (3), $\alpha _{lu}/n_{\mathrm{H}_2\,}$ is only a function of temperature and, following Clark et al. (2011a), we can express the optical depth by

Equation (5)

where

Equation (6)

defines an effective column density. Hence, the last task for the determination of optically thick cooling is the calculation of the characteristic length and the associated effective column density. Knowing these quantities, one can generally determine the angle-dependent escape probability and afterward average it over all lines of sight and all relevant ro-vibrational lines. Since H2 cooling is due to a number of lines without a single dominant line (Haardt et al. 2002), we can directly average the escape probability over these lines. Furthermore, one often assumes spherical symmetry, whereas Yoshida et al. (2006) proposed an average over three orthogonal directions

Equation (7)

The averaged escape probability finally relates the optically thin and optically thick cooling rate by

Equation (8)

and is therefore also known as the "opacity correction." In order to determine this escape probability, we have to understand the dynamics of the cloud. If the photon is emitted in the center of the cloud and an envelope of gas is moving toward it with a constant radial velocity gradient dvr/dr, then the photon observes the spectral lines of the envelope to be Doppler-shifted with respect to its rest frame. According to Sobolev (1947), a photon is not absorbed and can escape freely, if the spectral lines of a possibly absorbing H2 molecule are shifted by more than one thermal line width. The line width of thermal line broadening is given by

Equation (9)

where ν0 is the central frequency of the line, vth is the thermal velocity of molecular hydrogen, T is the temperature, and mH is the mass of a hydrogen atom. Using this, we can determine the characteristic distance Lchar beyond which the Sobolev criterion is fulfilled. This length scale is typically known as the Sobolev length

Equation (10)

Phrased differently, all relevant matter that might reabsorb a photon is within its Sobolev length. Assuming a constant density within this Sobolev length, the effective column density can be determined by

Equation (11)

In order to capture the three-dimensional dynamics of the collapse, one normally uses

Equation (12)

for the determination of the Sobolev length (Neufeld & Kaufmann 1993). However, a fundamental problem of the Sobolev approximation was already mentioned by several authors: both the velocity gradient and the number density have to be constant within one Sobolev length (Lucy 1971; Bujarrabal et al. 1980; Hummer & Rybicki 1992; Neufeld & Kaufmann 1993; Wolcott-Green et al. 2011). As we will see below, this assumption is generally not valid.

2.3. Correction of Sobolev Approximation

For simplicity, Zanstra (1934) assumed the absorption coefficient αlu to be zero outside the interval [ν0 − Δνth, ν0 + Δνth] and Sobolev (1947) used the same simplification. Following this approach, a photon can escape freely from the optically thick gas after one Sobolev length Ls, because it will not be reabsorbed thereafter. The actual absorption probability however, is related to the true overlap of spectral lines. In the present context, the shape of the H2 lines is dominated by thermal broadening and can therefore be described by a normalized Gaussian profile. If one line with central frequency ν0 is Doppler-shifted to the frequency ν, the relative line displacement in units of the thermal gas velocity between these two cases is given by

Equation (13)

Accordingly, the relative overlap of these two line profiles can be calculated by

Equation (14)

The overlap of spectral lines for the special case x = 1 is illustrated in Figure 1. This displacement represents the case after one Sobolev length, though the relative overlap is still 62%. While the original Sobolev approximation ignores all possible absorption events beyond this point, there is a non-negligible absorption possibility beyond one Sobolev length and we therefore need to introduce a correction term in order to account for the additional matter. Even after three Sobolev lengths the relative overlap, for example, is still 13% and reabsorption might be possible.

Figure 1.

Figure 1. Normalized thermal line profiles as a function of frequency in units of the thermal line width. The right profile is shifted by one thermal line width with respect to the left profile. The overlapping area shows a relative overlap of 62%, which should be negligible according to Sobolev (1947).

Standard image High-resolution image

In order to find a proper correction for this overlap, we start from the basic definition of the column density. However, we are not interested in the total column density but rather in the effective column density, which includes only the gas that could be relevant for the reabsorption of escaping H2 line photons. The Sobolev approximation gives a very simple answer to the question of which gas we have to include in the effective column density, namely all gas within one Sobolev length. Expressed in terms of line overlap we write this as

Equation (15)

where a constant number density of molecular hydrogen is assumed in the last step. Thus, Sobolev (1947) overestimated the overlap and the matter contribution within one Sobolev length but neglected all matter contributions beyond this point. Since we want to account for the true overlap of spectral lines, we introduce the relative line overlap o(x) as a weighting function into the determination of the effective column density,

Equation (16)

The remaining problem is to relate the line displacement x to the distance r along the line of sight. Assuming a constant velocity gradient dvr/dr, we can rewrite Equation (10) into

Equation (17)

and transform the radial integration into an integration over the relative line displacement

Equation (18)

which finally yields

Equation (19)

In other words, the relevant column density for the determination of the escape probability of line photons is about 1.7 times higher than originally assumed by Sobolev. Although we correct for the line overlap, we should still keep in mind that this derivation implies several assumptions like constant velocity gradient and constant density within the Sobolev length. Also some gas further away may be absorbing, because of similar velocities. In particular in turbulent clouds this may become relevant. This is related to the question of how "real" clumps identified in position–position–velocity space are in position–position–position space (see, e.g., Ballesteros-Paredes & Mac Low 2002; Beaumont et al. 2013).

Molecular hydrogen has more than two hundred spectral lines in the range 1.0 μm ⩽ λ ⩽ 32 μm which could be relevant for cooling primordial gas (Ripamonti et al. 2002). Although these lines appear to be very close to each other, the possibility that a photon is emitted in one line with a certain frequency and absorbed by another line which is Doppler-shifted into the emitting frequency is negligibly small (∼10−6). Consequently, this effect can be ignored.

2.4. Further Cooling Approaches

Besides the commonly used Sobolev approximation there are several other approaches, which we present in the following sections.

2.4.1. Gnedin Approximation

Gnedin et al. (2009) model molecular hydrogen and star formation in cosmological simulations. They determine H2 column densities for the self-shielding of H2 against Lyman–Werner photons by using a "Sobolev-like" approximation. They define a characteristic length scale

Equation (20)

based on the number density and its gradient. The column density is then simply given by

Equation (21)

This approach accounts for the density gradient in the gas distribution. In the following, we will label this method as "Gnedin." They claim that this approximation provides a very good estimate for the column density in the range $3 \times 10^{20} \mathrm{cm}^{-2} < N_{\mathrm{HI}}+2N_{\mathrm{H}_2\,} < 3 \times 10^{23}\, \mathrm{cm}^{-2}$. Although we are dealing with star formation on much smaller scales than they do, their method seems to be a reasonable, scale-invariant approach for the determination of effective column densities. Furthermore, this method depends only on local quantities and is easy to implement in different codes. Nevertheless, this method includes no information about the velocity profile of the cloud and therefore neglects the enhanced photon escape probability due to the Doppler-shifting of lines.

2.4.2. Analytic Fit Functions

We are interested in the column densities in order to determine the photon escape probabilities for H2 line cooling (Equation (8)). Besides the previously presented methods, there are two analytical fitting functions that directly relate a given number density of the gas to the photon escape probability. The first fitting function

Equation (22)

with nRA = 8 × 109 cm−3 and bRA = 0.45 was proposed by Ripamonti & Abel (2004) and has been applied in several (mainly grid-based) simulations (O'Shea & Norman 2006; Turk et al. 2011; Hirano & Yoshida 2013; Greif et al. 2014a). This formula was obtained from the detailed one-dimensional calculation by Ripamonti et al. (2002). A second method was proposed by Greif et al. (2013) who study the chemo-thermal instability in primordial star-forming clouds. The idea follows the approach by Ripamonti & Abel (2004) but with a smooth transition and therefore a continuous derivative toward the optically thin regime. The formula is given by

Equation (23)

where x = n/nG, nG = 4 × 109 cm−3, and bG = 0.45. These fits are the easiest and most direct way to determine the photon escape probability, but their use does not account for any information about the temperature, velocity or density profiles. Recently, Greif et al. (2014a) investigated the collapse of primordial gas using a multi-line, multi-frequency ray-tracing scheme in order to accurately model the transfer of H2 line emission. Although he provides a new fit formula, we will not include this in our analysis, since it is very similar to the original fit function by Ripamonti & Abel (2004).

2.4.3. Reciprocal Approach

In order to find a method for the determination of optically thick cooling that only depends on local quantities of the collapse and that is easy to implement, but nevertheless captures the dynamics of the cloud, we propose a combination of the Sobolev and Gnedin approximation. The corrected Sobolev approximation takes the line overlap into account but neglects the decreasing density. On the other hand, the Gnedin approximation takes the decreasing number density into account but neglects the Doppler-shifting of lines. Since each method on its own generally overestimates column densities, the general idea behind this new approach is to combine these two methods in order to overcome their individual shortcomings. The reciprocal sum

Equation (24)

of the two characteristic lengths provides the right behavior. If both lengths are of the same order (L1  ≃  L2), the result should be smaller than both lengths (Lrec  <  L1  ≃  L2), since each length individually overestimates the column density. Whereas if one length is significantly smaller than the other length L1  ≪  L2, the result should be equal to the smaller one (LrecL1), because beyond this smaller distance the photons can escape freely anyway. Following this method, the number density is simply given by

Equation (25)

3. NUMERICAL METHODS

Here, we describe the implementation of TreeCol and three criteria to define the effective column density. Furthermore, we present our simulations and the initial conditions.

3.1. Effective Column Densities with TreeCol

The main problem of the Sobolev approximation is the dependence on local quantities and therefore the neglect of all information about density gradients, velocity profile and the actual shape of the cloud. Generally, the most exact way to determine effective column densities is to sum up all relevant mass along all possible lines of sight. In order to avoid this extremely high computational effort, Clark et al. (2012) designed the TreeCol algorithm, which determines column densities based on a tree structure, used by many gravitational N-body solvers. TreeCol uses a spherical pixelation with diamond-shaped pixels based on HEALPix (Górski et al. 2005). During the walk of the tree, all relevant data for the column density map are collected and projected onto a spherical grid. Since the data are already stored in a tree, TreeCol can use this information and therefore scales as Nlog N with the number of cells or particles N. However, in our simulations, the usage of TreeCol slows down the simulation by a factor of about five with respect to a run without TreeCol. This slowdown is mainly related to the evaluation of several inverse trigonometric functions. Although we use this method in an smoothed particle hydrodynamics (SPH)-based simulation, the only requirement for its implementation is the clustering of matter in a tree-like structure, as indeed, the TreeCol method has already been implemented in the Arepo moving mesh code (Smith et al. 2014) and the FLASH AMR code (R. Wünsch et al., in preparation).

TreeCol overcomes several shortcomings of the Sobolev approximation. First of all, we can use it for any density distribution because we directly sum up the individual mass contributions. Furthermore, we can use the velocity information of the tree nodes and do not have to assume a constant velocity gradient. Additionally, we account for the actual spatial matter distribution and do not have to stick to a rough one dimensional approximation. While the original TreeCol algorithm computes the H2 column densities of the entire cloud, in what follows, we will present three improved versions that account for the Doppler-shifting of the line by only including mass that lies within the appropriate velocity range.

3.1.1. Sobolev-like

The Sobolev approximation assumes that all relevant mass for the column density is located within one Sobolev length. Translated into velocities, we should only include particles or tree nodes whose relative velocity is smaller than the thermal velocity. Thus, we modify TreeCol in order to include only the mass contributions of nodes that fulfill this criterion. A schematic illustration of this approach is given in Figure 2. From there we already see that the relevant volume around a particle does not necessarily have to be spherical, but usually follows the dynamic shape of the cloud.

Figure 2.

Figure 2. Two-dimensional illustration of three different approaches for the determination of effective column densities. Top left: original Sobolev method without TreeCol. Top right: Sobolev-like method implemented in TreeCol. Bottom: lookup method implemented in TreeCol. The circles represent individual mass contributions from cells or particles and the squares represent all mass in this tree node, which are clustered as seen by the black particle. The arrows indicate the relative velocities with respect to the black target particle, for which the column density should be determined. The original, isotropic Sobolev method defines a characteristic length scale, within which all matter contributes to the effective column density. In order to affect the effective column densities of the target particle in the Sobolev-like implementation, the relative velocities of the tree nodes have to be smaller than the thermal velocity vth of the target particle. In the lookup implementation, the matter contributions to the column density are weighted by the relative overlap of spectral lines, which in turn depends on the relative velocities. This weighting is illustrated by the partial filling.

Standard image High-resolution image

3.1.2. Corrected Sobolev

Following Equation (19), the next possible approach is to include all tree nodes whose relative velocities fulfill the criterion |vr|  <  1.694vth. Using this criterion, we take into account the non-negligible overlap beyond one Sobolev length. However, the individual velocity contributions are still not taken into account correctly, and we will focus on the following, more sophisticated approach for further studies.

3.1.3. Lookup Method

The above methods only distinguish whether matter contributes to the effective column density or not. However, the Doppler-shifting of spectral lines is a smooth process and so is the matter contribution, which should be included in a proper treatment of line cooling. The overlap of spectral lines and therefore the relevance for the effective column density depends only on the relative velocity. Since we know this information for the individual tree nodes, we weight their contributions to the column density map with their relative spectral line overlap (Equation (14)). A schematic illustration of this approach is given in Figure 2. This method does not rely on any assumptions (like a constant velocity gradient or number density), but captures the complete three-dimensional collapse of the cloud and takes care of the true line overlaps. This is the most exact method and it will serve as a reference for our analysis. For simplicity, we will refer to this method simply as "TreeCol," but actually mean the lookup method implemented in TreeCol—if not explicitly stated otherwise.

3.2. Simulations

In this section, we describe the general properties of our code, discuss the implementation of the TreeCol method, comment on its computational effort, and present our initial conditions.

3.2.1. Methods

We use a modified version of the cosmological simulation code Gadget-2. The original code was written by Springel (2005) in order to simulate structure formation by means of SPH. For the nonlinear collapse, we follow the chemical and thermal evolution of the gas cloud, solve the rate equations and couple the relevant heating and cooling terms to the hydrodynamic equations. For this purpose, we use a chemical network and cooling functions developed by Glover and collaborators (see Glover & Jappsen 2007; Glover & Abel 2008; Clark et al. 2011b and Clark et al. 2011a for full details). The primordial chemistry network includes H, D, He, H2, H+, H, D+, H$_2^+$, HD, He+, He++ and e. The rate equations between these species are solved self-consistently for every time step. For the three-body H2 formation reaction

Equation (26)

we adopt the rate coefficient of Glover (2008), derived by applying the principle of thermal balance to the collisional dissociation rate of Martin et al. (1996). The time scales associated with chemical heating and cooling can become extremely short when species come close to chemical equilibrium. Since the particle time step depends on the thermal timescale, this can result in some particles having very short timescales once the gas density exceeds ∼1017 cm−3. At this point it can become computationally prohibitive to run the simulation further.

Besides H2 line cooling, which is the most relevant process for this analysis, we also track the contribution by several other heating and cooling processes, such as electronic excitation of H and He, recombination, photodissociation, HD line cooling, Compton cooling, and bremsstrahlung (see Ripamonti & Abel 2004; Glover & Jappsen 2007; Clark et al. 2011a for a more detailed description). For the collisional induced emission (CIE) cooling, we follow the implementation by Clark et al. (2011b), which is based on studies by Ripamonti et al. (2002), Ripamonti & Abel (2004), and Yoshida et al. (2008). Above densities of nCIE ≈ 1014 cm−3, CIE cooling from H2 becomes more efficient than H2 line cooling, but the gas is most susceptible to fragmentation at densities below nCIE (see Section 6.2). Consequently, the implementation of CIE cooling is unlikely to significantly affect the conclusions of our study.

Furthermore, we implement a simple feedback model for the accretion heating by assuming that the accreting protostars produce an accretion luminosity of $L_\mathrm{acc} = G M_{*} \dot{M}/R_{*}$, where $\dot{M}$ is the mass accretion rate and M* and R* are the mass and radius of the protostars, respectively. We use the models by Smith et al. (2011) for the protostellar radius and assume a constant accretion rate of $\dot{M}\,{=}\,10^{-2}\, M_\odot\, \mathrm{yr}^{-1}$, which is consistent with Clark et al. (2011b), Hirano & Yoshida (2013), and with the actual accretion rates in our model. The heating rate is then given by

Equation (27)

where ρ is the gas density, κP is the Planck mean opacity and r is the distance from the source.

In order to minimize computational effort in the high-density regime, we use sink particles, based on the implementation by Jappsen et al. (2005), which was already used before (e.g., Boss & Black 1982; Boss 1987, 1989; Bate et al. 1995). Above a certain density threshold ncrit all SPH particles are merged into one single sink particle, which now contains all mass and momenta of the merged particles. This approach conserves mass and momentum, avoids small dynamical and chemical time steps, and we can identify the protostars by the newly formed sink particles. The critical density threshold is chosen so that the local Jeans mass is resolved by at least 100 SPH particles, following the resolution criterion by Bate & Burkert (1997).

Nevertheless, one should keep in mind that sink particles are not physical entities in their own right, but rather computationally motivated and consequently may cause problems by introducing a discontinuity in the mass and number of particles as well as a lack of pressure forces at the accretion radius. Strictly speaking, they also violate the hydrodynamic equations because the accretion onto the sink particle happens instantaneously (Greif et al. 2012). In order to guarantee that the formation of a sink particle actually represents the local collapse to a protostar, we introduce several formation criteria. The gas clump converted into a sink particle must exceed a certain density threshold, it must be gravitationally unstable, at the center of a locally convergent flow, and it must have a certain distance to preexisting sink particles (see, e.g., Federrath et al. 2010).

Another crucial value is the accretion radius racc, because a too large accretion radius might artificially influence the fragmentation behavior (Clark et al. 2011b; Greif et al. 2011; Machida & Doi 2013). On the other hand, Greif et al. (2012) and Machida & Doi (2013) state that fragmentation does not occur for densities above n ≃ 1017 cm−3, because there are no more efficient chemical or radiative cooling mechanisms. Smith et al. (2011), who additionally include the effect of heating by accretion feedback, find this value to be n ≃ 1015 cm−3. Phrased differently, the choice of the accretion radius might have an influence on the fragmentation behavior but as long as the critical density is ncrit  ≳  1016 cm−3, we should capture all fragmentation of the cloud. We set the critical density to ncrit  =  nres  =  3.42 × 1015 cm−3 in order to resolve the Jeans mass throughout the simulation. The Jeans length is λJ = 0.3AU under these conditions. Such a small accretion radius might lead to tiny dynamical time steps in the vicinity of sink particles and hence increase the computational effort. Therefore, we set the accretion radius to racc = 10 AU, which clearly fulfills the resolution criterion, but may suppress small-scale fragmentation.

3.2.2. Implementation of Cooling Approaches

The general idea of the different cooling approaches has already been discussed in Section 3.1. Here, we address the actual implementation of these approaches into the code. The optically thick cooling rate $\Lambda _{\mathrm{H}_2\,,\mathrm{thick}}$ is given by Equation (8), where the photon escape probability β can be expressed as a function of the column density divided by the thermal velocity of the H2 molecules (N/vth) and the temperature T. These values are stored in a lookup table for 31.6 K ⩽ T ⩽ 31, 600 K and 1017s  cm−3N/vth ⩽ 1027s  cm−3.

For the TreeCol-based determination of the effective column densities we create another lookup table, which relates the relative velocities in units of the thermal velocities (vr/vth) to an overlap of spectral lines. Since the relative velocities are distributed roughly equally throughout the simulation, we create this lookup table with linear steps in velocity space. For each node that might contribute to the effective column density we first check if

Equation (28)

because otherwise the line overlap is smaller than 10−3 and can be neglected anyway. The computational effort for the lookup of relative overlaps is negligibly small compared to the computational cost of TreeCol itself.

3.2.3. Initial Conditions

The initial conditions of our simulations presented here are generated in the following way: First, we ran collisionless N-body simulations with 1 Mpc/h comoving in length with uniform mass resolution to capture collapsing dark matter halos starting at a redshift z = 100. These collisionless simulations were executed with Gadget-3 (Springel et al. 2005) and we employ cosmological parameters consistent with the WMAP-7 measurements (Ωm = 0.271, ΩΛ = 0.729, σ8 = 0.809, h = 0.703; Komatsu et al. 2011). We would not expect our results to differ significantly if we were to use the cosmological parameters measured by Planck (Planck Collaboration 2014). This choice of parameters is for consistency with our previous works (Sasaki et al. 2014). Since we ensure that the parameters for the tree force calculation are set to the same value in the Gadget-3 and Arepo simulations, described later in this section, this is practically equivalent to using Arepo for the N-body simulation. The reason we used Gadget-3 to evolve dark matter only initial conditions is that, in principle, it allows us to make use of more physical merger histories of dark matter halos when picking a specific dark matter halo to resimulate. Currently, this is only analyzed with ease from Gadget-3 snapshots but not from Arepo snapshots. From this parent simulation, we select four dark matter halos, which are individually shifted to the center of the simulation box. The regions further away from each specific halo are represented by dark matter particles of proceedingly larger mass and lower resolution. The parent simulation contains 2563 (halo 1) and 5123 (halo 2–4) particles. At the regions of interest the mass resolution of the dark matter particles is improved by 43. The dark matter mass is 7.3 M for the 5123 runs.

From this newly generated initial condition with improved dark matter resolution, we start a hydrodynamical simulation with the moving-mesh code Arepo (Springel 2010). At the start-up of our hydrodynamic simulations, the code generates gas cells from dark matter only initial conditions. In order to numerically follow the dynamics of pristine gas in a cosmological context, we adopt an on-the-fly mass refinement scheme, which ensures that the Jeans length is resolved by at least 64 cells at all times. In order to deal with memory consumption and small time steps, we stop our simulation when the highest density in the simulation is ∼107 cm−3.

We cut out spheres with a radius of R = 0.5 pc, store only gas cells, and continue the simulation forcing further mass refinement in a non-cosmological setup (256 cells per Jeans length). This is to ensure that the Jeans mass at n ∼ 1017 cm−3 is resolved with ∼100 cells without further refinement in the Gadget-2 simulation.

The resulting gas clouds serve as initial conditions for the high-resolution runs, which are performed with Gadget. These clouds have average masses of ∼1.0 × 103M, temperatures of ∼400 K, and a mean H2 abundance of $x_{\mathrm{H}_2} = 1.1 \times 10^{-3}$. The average particle number is ∼2 × 107 per simulation, which yields a mass resolution of 10−3M in the central region. These final runs are performed in Gadget-2, to make use of the already implemented primordial chemistry.

An additional external pressure term is added in order to compensate for the missing gas contribution from the surrounding halo (Benz 1990; Clark et al. 2011a). In order not to artificially squeeze the cloud, we set the external pressure to the smallest occurring pressure in the outer 10% of the cloud.

4. RESULTS

In this section, we first test the validity of our new method in a simple, spherically symmetric test scenario. Then, we compare the TreeCol-based cooling approach to the local, isotropic cooling implementations for the cosmological halos.

4.1. Test Scenario

Before we apply our new method to cosmological halos, we test its accuracy in a simple test scenario. We start from a spherical, primordial cloud with a mass of 103M, an initial density of ∼104 cm−3 and a temperature of ∼250 K. The gas is represented by 643 SPH particles and we follow the simulation to densities above ∼1013 cm−3, where we insert sink particles. For each different method, we run the simulation independently and compare the effective H2 column densities in Figure 3. The lookup approach explicitly accounts for the true line overlaps and therefore is the most exact method. While "TreeCol corrected Sobolev" accounts for the correction factor and reproduces "TreeCol lookup" remarkably well, "TreeCol Sobolev" corresponds to the uncorrected Sobolev method and generally underestimates the effective column density. Consequently, this correction factor is mandatory to account for the exact line overlaps.

Figure 3.

Figure 3. Effective H2 column density as a function of density for different cooling approaches. Here and in the following figures (if not stated otherwise), the plotted values represent the means of the density-binned SPH particles. Staring from spherically symmetric initial conditions, this snapshot is taken immediately after formation of the first protostar. The three TreeCol-based methods refer to the different implementations of this algorithm (Section 3.1) and the Sobolev and Gnedin method are only based on local quantities. For the relevant, optically thick densities above ∼1010 cm−3, the local methods generally yield higher values than the more accurate TreeCol methods.

Standard image High-resolution image

Furthermore, the commonly used Sobolev method and the Gnedin approach, which are only based on local gas quantities, overestimate the column density in the optically thick regime above ∼1010 cm−3. These two local methods yield almost the same results for the effective column density, which means that the relation $v_\mathrm{th}/|\mathbf {\nabla }\cdot {\mathbf {v}}| \approx n_{\mathrm{H}_2\,}/|\mathbf {\nabla } n_{\mathrm{H}_2\,}|$ holds for the spherical collapse scenario. This illustrates that the density is generally not constant within the Sobolev length, but rather varies on the same length scale as the velocity field. Hence, even in this simple test case, the local methods are not able to determine the column density properly.

4.2. Comparison of Cooling Approaches

In this section, we test our approaches in more realistic cosmological halos. We first use the TreeCol-based simulations and determine all relevant information by post-processing these output files. By doing so, we can focus on the intrinsic differences of the methods (determined under the same physical conditions) rather than comparing different simulations with presumably different dynamics. The plots in this section represent averages over all four cosmological halos.

We want to find an accurate cooling approach that reproduces the TreeCol method best and hence, we compare the commonly used Sobolev approximation with the corrected Sobolev approximation, the Gnedin approach and our new reciprocal method, which combines the corrected Sobolev and Gnedin approach. The column density for these different approaches can be seen in Figure 4. Generally, the local methods overestimate the effective column density in the optically thick regime (n ≳ 1010 cm−3). Especially at later stages of the collapse these differences increase, because the slope of the photon escape fraction as a function of density flattens with time in the optically thick regime for the TreeCol method. The (corrected) Sobolev method overestimates the column density all the time, whereas the Gnedin and reciprocal approaches overestimate the column density only for high densities. The latter yield accurate fits for the density regime 109 cm−3n ≲ 1012 cm−3, whereby one should keep in mind that the relative importance of H2 line cooling decreases above n ≃ 1013 cm−3. Realizing that the Sobolev approximation already overestimates column densities, one might ask why we need this additional correction factor, which makes the approximation even worse. A detailed answer to this question is given in Section 6.1, but we already want to emphasize the total neglect of any density gradient. The Sobolev approximation assumes a constant density, although the density of molecular hydrogen generally decreases when moving radially outward. Hence, this approximation is not valid, but leads to an overestimation of the column density (already for the uncorrected Sobolev method).

Figure 4.

Figure 4. Effective H2 column density as a function of nH, the number density of H nuclei, for different cooling approaches. These snapshots are taken immediately before formation of the first sink particle (top) and immediately after formation of the second sink particle (bottom). "Reciprocal" represents the results based on the reciprocal sum of the Gnedin and the corrected Sobolev length. The solid red line indicates the TreeCol approach, which should be the most accurate of our models and the black dot-dashed line represents the H2 column density above which the gas is optically thick, computed assuming a fixed temperature T = 1000 K. Almost all methods tend to overestimate the effective column density in the optically thick regime, some by up to two orders of magnitude. The TreeCol column density decreases with time, while the other methods yield even higher column densities in the high-density regime.

Standard image High-resolution image

The photon escape probability as a function of density can be seen in Figure 5 for different methods. Although there are slight differences between the individual approaches, all methods seem to agree well with the TreeCol approach prior to the formation of the first sink particles. However, at later times, the TreeCol approach yields higher values for the photon escape probability, corresponding to a smaller effective opacity of the cloud. The other methods, which depend only on local quantities, cannot reproduce this behavior and therefore underestimate the photon escape probability.

Figure 5.

Figure 5. Photon escape probability as a function of density for different cooling approaches. These snapshots are taken immediately before formation of the first sink particle (top) and immediately after formation of the second sink particle (bottom). The solid red line indicates the TreeCol approach, which should be fitted by the other methods. While all the methods provide a good fit to the TreeCol method at low densities, the slope of the TreeCol approach flattens for later stages and the other methods cannot reproduce this behavior.

Standard image High-resolution image

We compare the analytical fitting formulae to the TreeCol method by using the parameterized formulae (Equations (22) and 23)) with the fit parameters nRA, bRA, nG, and bG. In Figure 6 we see the original functions and the best fits to the TreeCol data. The newly adjusted fits minimize the weighted scatter sum

Equation (29)

where NSPH is the number of particles, βanalytic(i) is the analytic fit, βTreeCol(i) is the photon escape probability based on the TreeCol approach, $\Lambda _{\mathrm{H}_2} (i)$ is the H2 cooling rate, and ΛCIE(i) is the CIE cooling rate for the ith particle, respectively. The weighting by the relative cooling rate accounts for the decreasing relevance of H2 cooling at higher densities.

Figure 6.

Figure 6. Photon escape probability as a function of density for two analytical fitting functions (the green dashed line shows the relative strength of H2 cooling). The snapshots are taken before formation of the first sink particle. The solid red line indicates the TreeCol approach, which should be fitted by the analytic formulae. The blue lines represent the original fits by Ripamonti & Abel (2004) and Greif et al. (2013), whereas the purple lines are their fit functions with the parameters adjusted to match our TreeCol method. Although the original fits are not reproduced, the data can generally be fitted by an analytic function for each specific time.

Standard image High-resolution image

For each single snapshot, the exact data can be fitted remarkable well by an analytic formula, but we note that both fit parameters vary strongly during the collapse. Although the original fits are a satisfying approximation at the beginning of the collapse, their accuracy decreases during the collapse and they totally miss the true photon escape probabilities at later stages.

The time evolution of the fit parameter b (Equations (22) and 23)) reveals an interesting insight into the structure of the collapse. In Figure 7 we compare the evolution of b for both fit formulae with time. It quantifies the slope and therefore represents the dependence of the opacity on density. The parameters b and n0 are fitted simultaneously, but the plot only contains the slope because it reveals more information about the underlying physics. For large b, the cloud becomes opaque with increasing density very promptly, whereas for a shallow slope the opacity remains low although the density increases. The blue dash-dotted line illustrates the original slope for both fits of b = 0.45, but the newly fitted formulae to the TreeCol data reveal that the actual slope is shallower and most notably varies with time. The small value of b for t < −500 yr is a numerical artifact, because at these times there are almost no particles in the optically thick regime, which might define a distinct slope. On the other hand, the shallow slope of 0.2 ⩽ b ⩽ 0.4 during the later stages of the collapse can be related to the flattening of the cloud. Consequently, it is impossible to find one single analytic fit, which describes the dynamics of the collapse completely. However, the mean parameters that fit the TreeCol data best from 500 yr before formation of the first protostar until the end of our simulations are nG = 5 × 109 cm−3 and bG = 0.32.

Figure 7.

Figure 7. Time evolution of the fit parameters bRA (Equation (22)) and bG (Equation (23)) to the TreeCol data (formation of the first sink particle at t = 0, averaged over four halos). The blue dash-dotted line illustrates the original slope of the fits (b = 0.45). Generally the newly fitted slopes are shallower than the originally proposed value. However, since variations in the slope are of chaotic nature, it is impossible to find one global fit function.

Standard image High-resolution image

In order to analyze the accuracy of the different cooling approaches quantitatively, we determine the relative error of the photon escape probability for each method. Therefore, we compare the photon escape probabilities of all particles above a certain density threshold to the ones determined with the TreeCol method. The density threshold is necessary, because below n = 109 cm−3, the escape probability is very close to one anyway and consequently there are no significant deviations between the methods. The relative error |βmethod − βTreeCol|/βTreeCol is weighted by the relative H2 cooling rate, so that the mean relative error of the photon escape probability can be expressed as

Equation (30)

where Nthresh is the number of particles above a certain density threshold. We chose the density threshold 109 cm−3 to analyze how accurate the methods are in this density regime. The time evolution of the mean relative photon escape probability error for this threshold can be seen in Figure 8. The relative error is small at early times in the collapse because all photon escape probabilities are close to one, whereas at later times, more particles enter the optically thick regime and the relative errors rise to values between 5%  and 30%. The reciprocal approach is slightly better than the Gnedin approximation, although both yield errors between 5%  and 15% throughout the simulations. Interestingly, the corrected Sobolev approximation yields higher errors (15%–30%), whereas the uncorrected one yields errors between 10%  and 25%. This behavior is caused by the fact that the Sobolev approximation already overestimates column densities because it neglects the density gradient. Since the reciprocal method is mostly dominated by the Gnedin approximation, we can conclude that the effective column density is more strongly influenced by the density gradient than by the relative velocities. However, the velocity information still improves the fit and is therefore important for a proper treatment. The original analytic fits are comparatively inaccurate with errors between 20%  and 50% and are therefore not displayed in this figure. However, the analytical fitting function proposed by Greif et al. (2013) yields slightly better results than the one proposed by Ripamonti & Abel (2004). The same analysis for higher density thresholds (1010 cm−3, 1011 cm−3) shows qualitatively similar results, but with a trend to higher mean relative errors.

Figure 8.

Figure 8. Mean relative error of the photon escape probability for different methods as a function of time (averaged over four halos, at t = 0 the first sink particle forms) for all particles above n > 109 cm−3. At early times, all photon escape probabilities are close to one and thus their mean error is small. At later stages, however, the relative error is between 5%  and 30% and the accuracy of the individual methods differ significantly. Data from all four simulations are combined in this plot.

Standard image High-resolution image

5. COLLAPSE AND FRAGMENTATION

Here, we analyze the fragmentation behavior of the gas with the different cooling implementations. We first compare the number of protostars that form in each simulation, analyze the dynamics of the collapse and then determine the susceptibility to fragmentation for the Sobolev method and the TreeCol approach.

5.1. Number of Protostars

We are interested in how the treatment of the cooling affects the characteristic mass range for Pop III stars. Based on our simulations, we compare the number of sink particles in each halo at the end of the simulations. However, due to a prohibitively small chemical time step and a too simplified treatment of feedback, we end the simulations before accretion has terminated. In order to compare the number of sink particles, we choose for each halo the snapshots with the same amount of accreted mass onto the sink particles. This condition enables us to compare the clouds for the Sobolev-based and the TreeCol-based runs at the same stage of the accretion process. The resulting numbers of sink particles are given in Table 1.

Table 1. Number of Sink Particles per Halo

Halo macc, tot Sinks Sobolev Sinks TreeCol
(M)
1 11.0 1 2
2 5.1 2 3
3 3.3 1 3
4 10.3 5 4

Notes. Comparison of the number of sink particles for four halos between the Sobolev approximation and the TreeCol approach. The same amount of accreted stellar mass guarantees comparability between the methods. In all but one simulation the TreeCol approach yields more sink particles.

Download table as:  ASCIITypeset image

The TreeCol approach yields more sink particles in all but one halo. However, the time evolution of the number of sink particles (see Figure 9) suggests that this effect might be caused by a delay of the fragmentation in the Sobolev-based runs. While the Sobolev approximation causes the disk to fragment into 1–6 protostars, the TreeCol approach causes slightly earlier fragmentation with at least two protostars in each halo. Since we had to end the simulations at this point, the final number of protostars cannot be fully constrained in this case. However, the TreeCol method accelerates the collapse and the gas fragments earlier compared to the Sobolev approximation. The accretion rates onto individuals protostars are in the range $3.3 \times 10^{-4}\, M_\odot\, \mathrm{yr}^{-1} \le \dot{M} \le 1.2 \times 10^{-2}\, M_\odot\, \mathrm{yr}^{-1}$ for the Sobolev approximation and $1.2 \times 10^{-3}\, M_\odot\, \mathrm{yr}^{-1} \le \dot{M} \le 1.7 \times 10^{-2}\, M_\odot\, \mathrm{yr}^{-1}$ for the TreeCol approach in all four halos.

Figure 9.

Figure 9. Number of sink particles as a function of time for the Sobolev-based runs (left) and the TreeCol-based runs (right). The clouds in the TreeCol approach fragment earlier, yielding more than two sink particles in each run, whereas the Sobolev approximation yields one to six sink particles per halo.

Standard image High-resolution image

5.2. Collapse Dynamics

In order to study the fragmentation behavior independently of the number of sinks, we first analyze the dynamics of the collapse. This analysis is mainly based on the fact that the temperature profile of a gas cloud directly influence the collapse to protostars by affecting the local Jeans mass, which is the absolute minimum requirement for collapse and is a strong function of temperature (Clark et al. 2011b). Hence, we compare the basic quantities between the four runs using the TreeCol approach and the four runs using the Sobolev approximation. The first quantity to compare is naturally the photon escape probability. Figure 10 illustrates the time evolution of the photon escape probability for both methods. While the spatial distribution of the photon escape probability seems to be smoother for the TreeCol-based run, the distribution is comparatively structured for the Sobolev approximation. Additionally, we see another important difference. Using the TreeCol method yields values for the photon escape probability close to one at the end of the simulations, whereas the Sobolev approximation yields smaller photon escape probabilities. Obviously, the Sobolev approximation is not able to capture the flattening of the cloud (see Section 6.1). Furthermore, the Sobolev-based simulation develops only one central core, while an elongated core with two peaks is formed in the TreeCol-based run. Since the photon escape probability has a direct influence on the cooling rates, we are consequently interested in the H2 cooling rate. This cooling rate is not significantly higher for the TreeCol-based run, although the photon escape probabilities are higher. This can be explained by the cooling implementation: due to the presumably higher cooling rate in the TreeCol-based run, thermal equilibrium is reached for lower temperatures. Thus, the cloud simply remains cooler instead of increasing its cooling rate significantly. Consequently, the gas in the Sobolev-based runs is generally hotter, which can also be seen in Figure 11.

Figure 10.

Figure 10. Mass-weighted projections of the photon escape probability in the central region for the Sobolev method (top) and the TreeCol method (bottom) at different times during the collapse: before sink formation (left), immediately after sink formation (middle), and ∼200 yr after the first sink particle formed (right). The photon escape probability appears to be smoother for the TreeCol approach and especially at late stages of the collapse the photon escape probability is close to unity for the TreeCol-based run. This increase in the photon escape probability is not reproduced by the Sobolev approximation. These plots are based on halo 1 but are representative for the other realizations.

Standard image High-resolution image
Figure 11.

Figure 11. Temperature as a function of density immediately after formation of the first sink particle for the four halos. Comparison of the Sobolev approximation (green dashed) and the TreeCol method (red solid). In the inner, high-density regime, the TreeCol approach yields lower temperatures by up to ∼500 K.

Standard image High-resolution image

The difference of temperatures in this inner regime can be up to ∼500 K, which influences the fragmentation behavior significantly.

5.3. Stability Analysis

Finally, we compare different fragmentation criteria of the clouds. The possibility of fragmentation in primordial clouds has been considered by many authors (e.g., Sabano & Yoshii 1977; Clark et al. 2008, 2011a, 2011b; Greif et al. 2011, 2012, 2013; Bromm 2013; Stacy & Bromm 2014). Fragmentation is a very chaotic, non-deterministic process and the actual outcome depends sensitively on the initial conditions (Girichidis et al. 2011, 2012a, 2012b). Nevertheless, there are three analytic expressions that help to quantify the possibility of a gas cloud fragmenting.

  • 1.  
    In order to locally contract instead of globally collapse, a necessary criterion for fragmentation of a gas cloud is (Rees & Ostriker 1977)
    Equation (31)
    where tcool = nkBT/(Λ[γ − 1]) is the cooling time with the net cooling rate Λ and $t_\mathrm{ff} = \sqrt{3 \pi /(32 G \rho)}$ is the free-fall time.
  • 2.  
    Toomre (1964) analyzed the stability of rotating gas disks and derives the instability criteria
    Equation (32)
    where κ is the epicyclic frequency and Σ is the surface density of the disk. Formally, this criterion is only valid for thin disks whereas Goldreich & Lynden-Bell (1965) extended the criterion by requiring Q < 0.676 for a finite-thickness isothermal disk to fragment.
  • 3.  
    Gammie (2001) investigated the nonlinear outcome of a stability analysis of a Keplerian accretion disk. Based on numerical experiments, he derived the instability criterion
    Equation (33)
    where ω is the orbital frequency. The Gammie criterion expresses the possibility that pieces of the disk cool and collapse before they have the opportunity to collide with one another in order to reheat the disk.

In their study on the formation and evolution of primordial protostellar systems, Greif et al. (2012) find the Toomre criterion insufficient for the quantification of gravitational instability and additionally use the Gammie criterion. None of these criteria guarantees fragmentation (Yoshida et al. 2006), but a combination of them yields a reliable quantification of instabilities in the gas disk.

We compare the radial and density profile of these fragmentation criteria for the Sobolev-based and the TreeCol-based runs. We do this directly after formation of the first protostar to ensure comparability between the different runs and methods (Figure 12). The requirement that all three fragmentation criteria have to be fulfilled simultaneously is very strict, because each criterion alone already quantifies stability. Applying this conservative criterion shows that all simulations are susceptible to fragmentations in the density regime 1010 cm−3  ⩽  n ⩽ 1011 cm−3 at a radius of ∼200 AU from the densest region. However, the Gammie criterion seems to be a less restrictive version of the Rees–Ostriker criterion and consequently yields no additional constraint to this regime, because it is fulfilled anyway. The regime of fragmentation and the instability of the gas cloud is slightly higher for the TreeCol approach. Especially the cooling time criterion remains significantly longer under the stability threshold for the TreeCol-based run, which is mainly a consequence of the lower temperatures, as discussed before. Consequently, the high-density gas is more susceptible to fragmentation in the TreeCol-based runs. Since these are averaged profiles over four halos, we expect these results to be representative for primordial star formation.

Figure 12.

Figure 12. Three different fragmentation criteria as a function of density (left) and radius (right) for the Sobolev-based (top) and the TreeCol-based (bottom) run. The data are averaged over four halos. Since these plots are qualitatively equal for all four halos, the average yields a representative quantification of fragmentation in primordial gas. These snapshots are taken between formation of the first and second protostar and quantify the susceptibility to disk fragmentation. We expect the disks to fragment if all three criteria are below one (purple dotted line). Although the requirement that all three fragmentation criteria have to be fulfilled simultaneously is a very strict one, all simulations are likely to fragment in the density regime 1010 cm−3n ⩽ 1011 cm−3 at a radius of ∼200 AU from the densest region (purple area). Since the temperature in this density regime is hardly affected by the choice of the cooling approach, this fragmentation regime is only slightly pronounced in the TreeCol-based run. However, at higher densities, tcool/tff remains longer in the instability regime and hence we expect the TreeCol-based runs to be more susceptible to fragmentation in the high-density regime because the gas can locally contract instead of globally collapse. Because of our sink accretion radius of 10 AU, the central region is artificially stabilized.

Standard image High-resolution image

6. DISCUSSION

In this section, we discuss the previous results, point out shortcomings of commonly used methods, and comment on the expected different fragmentation behavior.

6.1. Cooling Approaches

The local and isotropic approximations fail in determining the proper effective column densities for H2 cooling. Although they yield acceptable results up to the formation of the first protostars, most approaches break down during later phases of the collapse, because they cannot capture the evolving dynamical structure, particularly once a flattened accretion disk has formed. The appropriate consideration of the overlap of spectral lines is very important, although even the uncorrected Sobolev approximation generally overestimates column densities. Further shortcomings are mostly related to the assumption that all relevant quantities are constant within the characteristic length scale. This simplification avoids the evaluation of integrals along the line of sight but is formally only valid for large velocity gradients. The analysis of the relevant quantities shows that the H2 number density, thermal velocity and the gradient of the velocity vary by up to two orders of magnitude within one Sobolev length. While the thermal velocity seems to be rather constant, the H2 number density varies most strongly. Since the distribution around the central value is not symmetric, these effects do not cancel out and generally lead to an overestimation of column densities.

Due to angular momentum conservation in the infalling material, a disk forms around the first protostar and the photon escape probability is enhanced perpendicular to the disk. An isotropic column density estimation that depends only on local quantities, cannot capture this feature and therefore generally underestimates the photon escape probability (see Figure 13).

Figure 13.

Figure 13. Mass-weighted projections of the photon escape probability along the x- (left), y- (middle), and z-axis (right) for the Sobolev approximation (top) and the TreeCol method (bottom). This disk of halo 4, after several protostars have formed, is exemplary for the occurring processes. The flattened structure enables the photons to escape perpendicular to the disk. However, this enhanced photon escape probability is only reproduced by the TreeCol method because the Sobolev approximation is based on the assumption of spherical symmetry.

Standard image High-resolution image

The Sobolev approximation is not able to capture this angular dependence and consequently underpredicts the angle-averaged photon escape probability. Comparing the escape probabilities in different directions for one and the same particle quantitatively yields high angle-dependent variations. Shortly before formation of the first sink particle, the photon escape probabilities in the inner ∼100 AU vary by factors of up to 10–100 in different HEALPix pixels. During the further collapse, this value decreases but is still between 2  and 20 on average in the center of the cloud. This clearly demonstrates the failure of the local approximation in modeling inhomogeneous, non-spherical density distributions. Hirano & Yoshida (2013) also mention the problem of direction-dependent escape probabilities during their comparison of different cooling implementations in simulations of primordial star formation. Moreover, the Sobolev approximation is highly sensitive to small-scale details of the velocity field, and hence improving the resolution and accuracy with which one models the sub-sonic turbulence in the gas actually makes the method worse, because locally you start to see more of the turbulent velocity gradients, but these persist only over short length scales, not globally (Greif et al. 2014a).

Although the TreeCol-based method is computationally expensive, it is able to capture the non-uniformity of the H2 distribution in a way that local approximations cannot manage. It is also significantly cheaper than the more robust method employed by Greif et al. (2014a), although a comparison between the accuracy of the method proposed here and that by Greif et al. (2014a) still needs to be done.

Regarding the local methods, there is no general recommendation. Most of these methods assume spherical symmetry and hence their validity breaks down when a disk forms around the first protostar. Since the accretion process through the disk lasts for several thousand years or more, one should use an approach that yields a suitable long-time accuracy. The reciprocal method is the only approach that considers both gradients in density and the Doppler-shifting of spectral lines. Therefore, this method is the most accurate approach for the determination of the effective column density even after subsequent sink particle formation. The difference between whether one uses the corrected or uncorrected Sobolev length for the determination of the reciprocal sum is small, but the corrected Sobolev length provides a more accurate method for the long-time evolution of the disk. Analytic fits might be useful during the initial collapse of the primordial cloud, but also fail when the disk-like structure starts to form. Nevertheless, at any stage of the collapse, the escape probability as a function of mass can generally be fitted very accurately by an analytic formula, which could also be done on the fly during the simulation. Consequently, a hybrid version between the proper TreeCol-based determination of escape probabilities every several time steps and an analytic fit to these data for the steps in between should yield reliable results with less computational effort.

This new method for the determination of effective column densities can also be applied in other astrophysical scenarios. For example the photodissociation of H2 in protogalaxies is a crucial process for the formation of supermassive black hole seeds (Latif et al. 2013). To model these processes properly, the effective column densities are needed to account for the effect of H2 self-shielding against Lyman Werner photons (Wolcott-Green et al. 2011). It can generally be applied in most scenarios of line transfer, where the radially infalling velocities of gas are higher than the turbulent gas velocities.

6.2. Fragmentation

Fragmentation is a highly chaotic process and slight changes in the initial conditions or in the implementation of the governing physics can completely change the outcome. Therefore, looking just at the number of sink particles is not a valid quantification, especially since we do not capture the entire duration of the accretion and fragmentation process. Nevertheless, it is instructive to determine and compare different fragmentation criteria for the individual cooling approaches. We find that the TreeCol method seems to promote fragmentation. In other words, commonly used cooling approximations generally underestimate the number of Pop III stars. Moreover, the disk has just formed, when we end our simulations. Therefore, the differences between the individual cooling approaches and their effect on fragmentation might even be more pronounced for later stages of the collapse. Since the overall mass accretion rates are roughly equal, regardless of the cooling implementation, we expect the Pop III stars in our TreeCol-based runs generally to have smaller masses than previous Sobolev-based studies have yielded.

6.3. Caveats

There are several open questions, shortcomings and approximations, which one should keep in mind, when interpreting the previously presented results. Since we follow the fragmentation of the cloud only for the first few hundred years after formation of the first protostar, we have no information about the physical conditions during the late stages of disk accretion. It is, however, likely that the disk-like structure will proceed to grow and most of our statements remain valid.

Furthermore, we do not account for mergers, which influence the final number of Pop III stars (Greif et al. 2011, 2012; Stacy & Bromm 2014), nor do we include the effects of ionizing radiation. We determine radiative feedback under the assumption of a constant mass accretion rate and although the value of 10−2M yr−1 seems to be justified by similar accretion rates in the simulations, a proper treatment is necessary. Besides, we should also keep in mind that the escape probability, as the theoretical basis for our cooling implementation, is an approximation by itself, e.g., we use one average escape probability for all photons, instead of determining the individual escape probabilities for each line separately. In this context, Greif et al. (2014a) find that a multi-line, multi-frequency ray-tracing scheme does not alter these results significantly. Magnetic fields might influence the collapse and act as a stabilizing force. We have not included these effects in our simulation, but for a detailed discussion of the effects of magnetic fields in primordial star formation, see Machida & Doi (2013 and references therein).

7. CONCLUSION

We compared different implementations for the approximate treatment of optically-thick H2 line cooling and analyzed the fragmentation behavior of primordial gas under these different methods. Since H2 is the dominant coolant in primordial gas clouds, line cooling by molecular hydrogen is a crucial process to consider for the formation of Pop III stars. While the cooling rates in the optically thin regime can be calculated accurately, optically thick cooling is only poorly understood, although it has a strong influence on the temperature profile and fragmentation of the cloud. The commonly used Sobolev approximation has to be corrected for the effect of line overlap. However, since the Sobolev and other approximations for the effective column density assume isotropy and certain quantities to be constant on the relevant scales, they all fail if the system deviates strongly from spherical symmetry or if it has strong density gradients. While the cloud flattens and develops a disk during the collapse, the local approaches generally yield too small values for the photon escape probability (mean relative errors of ∼20%). Existing analytical fitting formulae yield acceptable results up to the formation of the first protostar. Thereafter, the detailed functional form needs to be modified, because an analytical fitting formula derived at one time in the evolution will generally fail at other times, as the system is strongly out of dynamical equilibrium and therefore rapidly evolving. Only the TreeCol-based methods are adaptively adjusting to the kinematic and morphological changes of the system. Capturing these dynamical features, the TreeCol-based methods yield lower temperatures in the center of the cloud.

We find primordial gas is most susceptible to fragmentation in the density regime from n = 1010 cm−3 to n = 1011 cm−3. Whereas local methods lead to the formation of fewer and higher-mass Pop III stars, the TreeCol-based approach promotes a higher degree of fragmentation and therefore tend to result in the formation of more and lower-mass stars. Regardless of the cooling implementation, the protostars have very high mass accretion rates and the mass function will be dominated by high-mass stars. However, only high-resolution simulations with a proper treatment of H2 cooling, magnetic fields, and feedback that run long enough to capture the entire accretion and fragmentation process can give a complete picture of Pop III star formation and the corresponding primordial IMF.

The authors acknowledge financial support from the Deutsche Forschungsgemeinschaft (DFG) via SFB 881 "The Milky Way System" (sub-projects B1, B2, and B8), and from the Baden-Württemberg-Stiftung by contract research via the programme Internationale Spitzenforschung II (grant P-LS-SPII/18). The simulations described in this paper were performed on the kolob cluster at the University of Heidelberg, which is funded in part by the DFG via Emmy-Noether grant BA 3706, and via a Frontier grant of Heidelberg University, sponsored by the German Excellence Initiative as well as the Baden-Württemberg Foundation. T.H. appreciates fruitful discussions with Clio Bertelli Motta and Volker Bromm. T.H. and R.S.K. are supported from the European Research Council under the European Communitys Seventh Framework Programme (FP7/2007-2013) via the ERC Advanced Grant "STARLIGHT: Formation of the First Stars" (project number 339177). M.S. and T.H. acknowledges computing time from JUDGE super computing center. P.C.C. is supported by grant CL 463/2-1, part of the DFG priority program 1573 "Physics of the Interstellar Medium." M.S., as part of IMPRS-HD, thanks financial support from Heidelberg Graduate School of Fundamental Physics (HGSFP) and wishes to thank Thomas Greif for help with performing numerical simulations.

Footnotes

Please wait… references are loading.
10.1088/0004-637X/799/2/114