-
PDF
- Split View
-
Views
-
Cite
Cite
Morag I. Scrimgeour, Tamara Davis, Chris Blake, J. Berian James, Gregory B. Poole, Lister Staveley-Smith, Sarah Brough, Matthew Colless, Carlos Contreras, Warrick Couch, Scott Croom, Darren Croton, Michael J. Drinkwater, Karl Forster, David Gilbank, Mike Gladders, Karl Glazebrook, Ben Jelliffe, Russell J. Jurek, I-hui Li, Barry Madore, D. Christopher Martin, Kevin Pimbblet, Michael Pracy, Rob Sharp, Emily Wisnioski, David Woods, Ted K. Wyder, H. K. C. Yee, The WiggleZ Dark Energy Survey: the transition to large-scale cosmic homogeneity, Monthly Notices of the Royal Astronomical Society, Volume 425, Issue 1, September 2012, Pages 116–134, https://doi.org/10.1111/j.1365-2966.2012.21402.x
- Share Icon Share
Abstract
We have made the largest volume measurement to date of the transition to large-scale homogeneity in the distribution of galaxies. We use the WiggleZ survey, a spectroscopic survey of over 200 000 blue galaxies in a cosmic volume of ∼1 h−3 Gpc3. A new method of defining the ‘homogeneity scale' is presented, which is more robust than methods previously used in the literature, and which can be easily compared between different surveys. Due to the large cosmic depth of WiggleZ (up to z = 1), we are able to make the first measurement of the transition to homogeneity over a range of cosmic epochs. The mean number of galaxies N(< r) in spheres of comoving radius r is proportional to r3 within 1 per cent, or equivalently the fractal dimension of the sample is within 1 per cent of D2 = 3, at radii larger than 71 ± 8 h−1Mpc at z ∼ 0.2, 70 ± 5 h−1 Mpc at z ∼ 0.4, 81 ± 5 h−1 Mpc at z ∼ 0.6 and 75 ± 4 h−1 Mpc at z ∼ 0.8. We demonstrate the robustness of our results against selection function effects, using a Λ cold dark matter (ΛCDM) N-body simulation and a suite of inhomogeneous fractal distributions. The results are in excellent agreement with both the ΛCDM N-body simulation and an analytical ΛCDM prediction. We can exclude a fractal distribution with fractal dimension below D2 = 2.97 on scales from ∼80 h−1 Mpc up to the largest scales probed by our measurement, ∼300 h−1 Mpc, at 99.99 per cent confidence.
1 Introduction
One of the main assumptions of the standard theory of cosmology, Λ cold dark matter (ΛCDM; based on CDM and a cosmological constant), is that the Universe is homogeneous and isotropic on large scales, and hence can be described by the Friedmann–Robertson–Walker (FRW) metric. ‘Homogeneous' means that its statistical properties (such as density) are translationally invariant; ‘isotropic' means it should be rotationally invariant. The Universe clearly deviates from this on small scales, where galaxies are clustered, but on large enough scales (≳100 h−1 Mpc in ΛCDM), the distribution of matter is assumed to be ‘statistically homogeneous' – i.e. the small-scale inhomogeneities can be considered as perturbations, which have a statistical distribution that is independent of position. However, this is merely an assumption, and it is important for it to be accurately verified by observation. Over the last decade, there has been a debate in the literature as to whether the Universe really is homogeneous, or whether it has a fractal-like structure extending to large scales. It is important to resolve this contention if we are to be justified in assuming the FRW metric.
In fact, although ΛCDM is based on the assumption of large-scale homogeneity and an FRW metric, inflation (which ΛCDM incorporates) actually predicts a certain level of density fluctuations on all scales. Inflation predicts that the primordial density power spectrum was close to scale invariant. In the standard model, the scalar index ns, which quantifies the scale dependence of the primordial power spectrum, is close to 0.96 (Baumann ), while a scale-invariant power spectrum has ns = 1. In this case, these density fluctuations induce fluctuations in the metric, δΦ, which are virtually independent of scale, and are of the order of δΦ/c2 ∼ 10−5 (Peacock ). Since these perturbations are small, the FRW metric is still valid, but it means that we expect the Universe to have a gradual approach to large-scale homogeneity rather than a sudden transition.
The most important implication of inhomogeneity is the so-called ‘averaging problem' in general relativity (GR). This arises when we measure ‘average' quantities (such as the correlation function and power spectrum, and parameters such as the Hubble constant) over a spatial volume. In doing so, we assume the volume is homogeneous and smooth, when it may not be. Since the Einstein equations are non-linear, density fluctuations can affect the evolution of the average properties of the volume – this is known as the ‘backreaction mechanism' (e.g. Buchert ; Ellis & Buchert ; Li & Schwarz ; see Räsänen , for a summary). If we observe a quantity within such a volume, we need to take averaging into account to compare it with theory. It is therefore important to know how much inhomogeneity is present, in order to obtain meaningful results from averaged measurements (and hence most, if not all, cosmological measurements).
Backreaction has also been proposed as an explanation of dark energy, which is believed to be a negative-pressure component of the Universe that drives the accelerated expansion. Some authors have suggested that instead of introducing exotic new forms of dark energy, or modifications to GR, we should revisit the fundamental assumptions of the ΛCDM model, such as homogeneity. If we assume that GR holds, but take inhomogeneities into account, it can be shown that backreaction can cause a global cosmic acceleration, without any additional dark energy component (see e.g. Schwarz ; Kolb et al. ; Räsänen , ; Wiltshire ; Buchert ). This effect appears to be too small to explain the observed acceleration, but highlights the importance of understanding the amount of inhomogeneity in the Universe.
Another, related, consequence of inhomogeneity is that it can affect the path travelled by light rays, and the calibration of clocks and rods of observers. It can therefore affect distance measurements, such as redshifts and luminosity distances (Wiltshire ; Meures & Bruni ). This has also been proposed as a possible explanation of the observed cosmic acceleration, although the effect appears to be of the order of only a few per cent at z ∼ 1 (Brouzakis, Tetradis & Tzavara ).
Homogeneity is required by several important statistical probes of cosmology, such as the galaxy power spectrum and n-point correlation functions, in order for them to be meaningful. Applying these to a galaxy sample below the scale of homogeneity would be problematic, since, if a distribution has no transition to homogeneity, it does not have a defined mean density, which is required to calculate and interpret these statistics. It is also not possible to model its cosmic variance, so the error in the measurements would be ill defined, making it impossible to relate these statistics to a theoretical model. It is therefore important to quantify the scale on which the Universe becomes close enough to homogeneous to justify their use.
Large-scale homogeneity is already well supported by a number of different observations. In particular, the high degree of isotropy of the cosmic microwave background (CMB; Fixsen et al. ) gives very strong support for large-scale homogeneity in the early Universe, at redshift z ∼ 1100. The isotropy of the CMB also indicates the Universe has remained homogeneous, since there are no significant integrated Sachs–Wolfe (ISW) effects distorting our view of the isotropic CMB (Wu, Lahav & Rees ). Other high-redshift evidence for homogeneity includes the isotropy of the X-ray background (Peebles ; Scharf et al. ), believed to be emitted by high-redshift sources, and the isotropy of radio sources at z ∼ 1 (Blake & Wall ).
However, these measurements of high-redshift isotropy do not necessarily imply homogeneity of the present Universe. If every point in the Universe is isotropic, then this implies the Universe is homogeneous; so if we accept the Copernican principle that our location is non-special, then the observed isotropy should imply homogeneity (Peacock ). However, most of these measurements (except the ISW effect) only tell us about the high-redshift Universe. We know that it has evolved to a clustered distribution since then, and it is possible that it could also have become anisotropic. It is also possible for the matter distribution to be homogeneous, while the galaxy distribution is not, since the galaxy distribution is biased relative to the matter field (Kaiser ) – although since galaxy bias is known to be linear on large scales (Coles ; Scherrer & Weinberg ), this seems unlikely. The ISW effect (Sachs & Wolfe ) gives information about the low-redshift Universe, since it mostly probes the dark energy dominated era, z ≲ 1 (Afshordi ), but it is an integral over the line of sight and so does not give full 3D information.
Galaxy surveys are the only 3D probe of homogeneity in the nearby Universe, and a number of homogeneity analyses have been carried out with different surveys, with seemingly conflicting results. Most statistical methods used to measure homogeneity have been based on the simple ‘counts-in-spheres' measurement, i.e. the number of galaxies N(< r) in spheres of radius r centred on galaxies, averaged over a large number of such spheres. This quantity scales in proportion to the volume (r3) for a homogeneous distribution, and homogeneity is said to be reached at the scale above which this holds. Hogg et al. () applied this to the Sloan Digital Sky Survey (SDSS) luminous red galaxy (LRG) sample at z ∼ 0.3, and found that the data became consistent with homogeneity at ∼70 h−1 Mpc for this galaxy population (using a different method of determining the homogeneity scale than we do).
This measurement can also be extended to a fractal analysis. Fractal dimensions can be used to quantify clustering; they quantify the scaling of different moments of galaxy counts-in-spheres, which in turn are related to the n-point correlation functions. The most commonly used is the correlation dimension D2(r), which quantifies the scaling of the two-point correlation function, and is based on the counts-in-spheres, which scale as . One can also consider the more general dimensions Dq, where q are different moments of the counts-in-spheres. Using fractal analyses, some researchers have found a transition from D2 < 3 to D2 = 3, at around 70–150 h−1 Mpc (Martínez & Coles ; Guzzo ; Martínez et al. ; Scaramella et al. ; Amendola & Palladino ; Pan & Coles ; Kurokawa, Morikawa & Mouri ; Yadav et al. ; Sarkar et al. ), whereas other authors have found no such transition (Coleman & Pietronero ; Pietronero, Montuori & Sylos Labini ; Sylos Labini, Montuori & Pietronero ; Joyce, Montuori & Sylos Labini ; Sylos Labini et al. ; Sylos Labini ). However, many of the galaxy redshift surveys used in the above-mentioned works are too shallow, sparse, or have survey geometries too complicated, to give conclusive results.
In this work, we use the WiggleZ Dark Energy Survey (Drinkwater et al. ) to make a new measurement of the counts-in-spheres and correlation dimension, to test for the transition to homogeneity. WiggleZ provides a larger volume than previous surveys, making it ideal for a homogeneity measurement, and it covers a higher redshift range, allowing us to also investigate how homogeneity changes with cosmic epoch. It is not volume limited, but we show that this does not significantly affect our measurement. The transition to homogeneity can be used as a test of a particular cosmological model, since we would expect it to differ for different cosmologies. In this work, we test a ΛCDM model with best-fitting parameters from the Wilkinson Microwave Anisotropy Probe (WMAP) data (Komatsu et al. ), which we refer to as ΛCDM+WMAP. We demonstrate the robustness of our measurement against systematic effects of the survey geometry and selection function, by repeating our analysis on both the GiggleZ N-body simulation and on a suite of inhomogeneous, fractal distributions.
Before we make any meaningful test of homogeneity, however, it is crucial to properly define what we mean by the so-called ‘scale' of homogeneity. Since there is only a gradual approach to homogeneity, such a definition may be arbitrary. In the past, authors have defined the ‘scale of homogeneity' as the scale where the data become consistent with homogeneity within 1σ (e.g. Hogg et al. ; Bagla, Yadav & Seshadri ; Yadav, Bagla & Khandai ). However, this method has several disadvantages. It depends on the size of the error bars on the data, and hence on the survey size. A larger survey should have smaller error bars, and so will automatically measure a larger scale of homogeneity. It also depends on the bin spacing, and is susceptible to noise between data points. We therefore introduce a different, and more robust, method for determining homogeneity: we fit a smooth, model-independent polynomial curve to all the data points, and find where this intercepts chosen values close to homogeneity.
Certain parts of our analysis require the assumption of a cosmological model and, implicitly, homogeneity (i.e. for converting WiggleZ redshifts to distances, correcting for the selection function, calculating the uncertainties using lognormal realizations and finding the best-fitting bias). In these cases, we use an input ΛCDM cosmology with h = 0.71, ,
,
, σ8 = 0.8 and ns = 0.96. Here, the Hubble constant is H0 = 100 h km s−1 Mpc−1, Ωm is the mass density,
is the dark energy density, Ωb is the baryon density, σ8 is the rms mass variation within spheres of 8 h−1 Mpc radius and ns is the spectral index of the primordial power spectrum. This is the same fiducial cosmology used by Blake et al. (), and we use this for consistency. We discuss the implications of assuming a ΛCDM model on the results of our homogeneity measurement in Section .
This paper is organized as follows. In Section , we describe the WiggleZ survey and our data set. In Section , we describe our methodology. We also explain our definition of the ‘scale of homogeneity' and present a new model-independent method for measuring this from data. In Section , we describe our analytic ΛCDM+WMAP model. We present our results in Section . We test the robustness of our method using fractal distributions and a ΛCDM N-body simulation in Section . We discuss our results in Section and conclude in Section .
2 THE WiggleZ SURVEY
The WiggleZ Dark Energy Survey (Drinkwater et al. ) is a large-scale spectroscopic galaxy redshift survey conducted at the 3.9-m Anglo-Australian Telescope, and was completed in 2011 January. It maps a cosmic volume of ∼1 Gpc3 up to redshift z = 1, and has obtained 239 000 redshifts for ultraviolet-selected emission-line galaxies with a median redshift of zmed = 0.6. Of these, 179 599 are in regions contiguous enough to be used for our analysis. It covers ∼1000 deg2 of equatorial sky in seven regions, shown in Drinkwater et al. (, their fig.1).
The observing strategy and galaxy selection criteria of the WiggleZ survey are described in Blake et al. () and Drinkwater et al. (). The selection function we use is described in Blake et al. (). The targets were selected from Galaxy Evolution Explorer (GALEX) satellite observations matched with ground-based optical measurements, and magnitude and colour cuts were applied to preferentially select blue, extremely luminous high-redshift star-forming galaxies with bright emission lines.
The WiggleZ survey offers several advantages for a new study of the scale of homogeneity. Its very large volume allows homogeneity to be probed on scales that have not previously been possible, and at a higher redshift; it probes a volume at z > 0.5 comparable to the SDSS LRG catalogue at z < 0.5. This allows us to make the first measurement of the change in the homogeneity scale over a range of cosmic epochs. We divide our sample into four redshift slices, 0.1 < z < 0.3, 0.3 < z < 0.5, 0.5 < z < 0.7 and 0.7 < z < 0.9. The sizes of the WiggleZ regions in each redshift slice are listed in Table 1; they sample scales well above the expected scale of homogeneity. The numbers of galaxies in each redshift slice are listed in Table 2.

Comoving dimensions of WiggleZ regions in each redshift slice, in units of h-1 Mpc. The dimensions correspond to the line of sight, RA and Dec. directions, respectively.

We also benefit from having seven regions distributed across the equatorial sky, which reduce the effect of cosmic variance. In addition, WiggleZ probes blue galaxies, whereas SDSS (which has obtained the largest scale measurements of homogeneity to date) used LRGs, and so it can also constrain any systematic effects introduced by the choice of tracer galaxy population (Drinkwater et al. ). Since blue galaxies are less biased, they are also more representative of the underlying matter distribution.
There are, however, several aspects of the survey that could potentially be detrimental to a homogeneity measurement. WiggleZ has a complex window function, with a complicated edge geometry including holes in the angular coverage, and the spectroscopic completeness varies across the sky. In addition, the population properties of the galaxies are known to vary with redshift, due to the effects of Malmquist bias (since WiggleZ is a flux-limited survey), downsizing (the observed fact that the size of the most actively star-forming galaxies decreases with time; Cowie et al. ; Glazebrook et al. ) and the colour and magnitude selection cuts (Blake et al. ). This means that WiggleZ preferentially selects larger mass, higher luminosity galaxies at higher redshift. A consequence of this is that it is not possible to define volume-limited subsamples of WiggleZ. However, we can correct for these effects by using random catalogues (Section ), which account for the survey selection function (Blake et al. ). We also divide the survey into four redshift slices, reducing the amount of galaxy population evolution in any region. In addition, we show that our results are not biased by the assumption of homogeneity in the selection function corrections, using an N-body simulation and a suite of inhomogeneous fractal distributions, described in Section . We are therefore confident that our result is not distorted by any features of the survey.


3 Methodology
Here we describe our methodology for measuring the transition to homogeneity. We first calculate the mean counts-in-spheres N( < r), then we find the fractal correlation dimension D2(r), from the slope of N(r). Although they are closely related, it is interesting to consider both, since the counts-in-spheres is the simplest measurement of homogeneity, whilst the correlation dimension provides direct information about the fractal properties of the distribution. We also describe our method of determining uncertainties, and our method of defining the ‘homogeneity scale' RH.
3.1 Scaled counts-in-spheres ![graphic]()





In each WiggleZ region, we make measurements in spheres with 12–15 logarithmically spaced radial bins (depending on the size of the region). We determine the large-scale cut-off in each region by calculating the mean volume in the selection function,
, in a thin shell of mean radius r enclosing a central galaxy. We illustrate this for the 15-hr 0.5 < z < 0.7 region in Fig. 1, and compare it with the ‘true' volume of the shells. Below ∼2 h−1 Mpc, there is noise due to low resolution, but above this the two curves are very close. On large scales, however, an increasing proportion of shells surrounding galaxies goes off the edge of the survey, and their volume within the survey decreases. On such scales, corrections for edge effects will become important. We take the large-scale cut-off of our homogeneity measurement at the radius of the maximum-volume shell. For the region in Fig. 1, this corresponds to shells that are ∼20 per cent complete. We show later that edge effects up to this scale do not impact our homogeneity measurement.

An illustration of the scales for which survey edge effects become important for the homogeneity measurement. Top panel: the mean volume in a thin shell of mean radius r surrounding a WiggleZ galaxy within the selection function (black curve). We show this for the 15-hr 0.5 < z < 0.7 region. The true volume of the shells is shown as a blue dashed line. The black curve deviates from the blue at large scales, where an increasing proportion of the shells extends outside the survey. Bottom panel: the ratio of the mean volume to the true volume, as a function of r. The grey dashed line indicates a ratio of 1.
Our correction method using random catalogues maximizes the use of the data, and accounts for incompleteness and the fact that WiggleZ is not volume limited. However, it can potentially bias our result towards detecting homogeneity, since it assumes homogeneity on the largest scales of the survey. It is equivalent to weighting each measurement by the volume of the sphere included within the survey, multiplied by an arbitrary mean density. Therefore, while the counts-in-spheres measurement should have the advantage of not assuming a mean density (Hogg et al. ), our correction method means that we do. Therefore, should tend to 1 on the largest scales of the survey, regardless of whether homogeneity has been reached. However, for a distribution with homogeneity size smaller than the survey,
should reach 1, and remain at 1, for a range of scales smaller than the survey scale. We check the robustness of our method against effects of the selection function and correction method in Section and show that our analysis is robust against this potential source of systematic error out to scales far greater than the homogeneity scale we measure.
There are other correction methods used in the literature. There is the so-called ‘exclusion' or ‘deflation' method (e.g. Coleman & Pietronero ; Pan & Coles , ; Sylos Labini et al. ), which only considers central points that are surrounded by complete spheres within the survey. This therefore excludes as central points any galaxies within a certain distance from the survey edges. However, this does not make the best use of the data, since it excludes data and so reduces the volume of the sample.
There is also the so-called ‘angular correction' model, which has been shown to be more optimal, by using all the available data without introducing a bias due to edge corrections (Pan & Coles ). This corrects measurements in a sphere by the solid angle subtended by regions in the sphere that are outside the survey boundary. However, as they point out, this method is difficult to apply to surveys with a complicated geometry, especially if they contain holes, as WiggleZ does. Finally, another correction that minimizes bias at the survey edges, but wastes little data, is the ‘Ripley' estimator (Ripley ; Martínez et al. ), which corrects the measurement in a sphere by the area of the sphere contained within the survey. Both the angular correction and the Ripley estimator assume isotropy of the samples. Due to the geometry of the WiggleZ survey, we choose to use a random catalogue correction, but make robustness tests to quantify any bias it may introduce to the results.
To demonstrate the robustness of our measurement against the method of correcting for the selection function, we compare our method to an analysis using only complete spheres, with and without correcting for incompleteness, in Section , and show we obtain consistent results.
3.2 Correlation dimension D2(r)
Fractal dimensions can be used to describe the clustering of a point distribution. There exists a general family of dimensions Dq, the Minkowski–Bouligand dimensions, which describe the scaling of counts-in-spheres centred on points (see, e.g., Borgani ; Martínez & Saar , for a review). To completely characterize the clustering of our galaxy distribution, we would need to consider all the moments q of the distribution (corresponding to combinations of n-point correlation functions). However, to identify the scale of homogeneity we consider only D2, the ‘correlation dimension', which quantifies the scaling behaviour of the two-point correlation function ξ(r).




For a homogeneous distribution, D2 = 3. If D2 < 3 then the distribution has a scale invariant, fractal (and so inhomogeneous) clustering pattern. If D2 > 3 the distribution is said to be ‘superhomogeneous' and corresponds to a lattice-like distribution (Gabrielli, Joyce & Sylos Labini ). A power law in 1 + ξ(r) ∼ r−γ has D2 = 3 − γ for ξ ≫ 1.
Some previous works have found D2(r) by simply fitting a straight line to a log–log plot of . This method can give a false indication of a fractal (Martínez et al. ); calculating D2(r) explicitly gives a more reliable measurement.





However, the correlation function cannot be used to test large-scale homogeneity, since the way it is determined from surveys depends on the mean galaxy density. Determinations of ξ(r) commonly use the Landy–Szalay (Landy & Szalay ) or Hamilton estimators (Hamilton ), which compare the galaxy clustering to that of random catalogues of the same mean density as the survey.
Our estimator also compares the data to random catalogues, since we must correct for the selection function. Therefore,
, like ξ(r), does assume a mean density on the scale of the survey. However, our estimator is slightly different, as we correct each object separately rather than an average pair count.
The correlation dimension D2(r), on the other hand, measures the scaling of , which is not affected by the assumption of the mean density. [This only affects the amplitude of
.] Deviations from a volume-limited sample, which require random catalogue corrections, only cause second-order changes to D2(r), whilst they would be leading order in the raw correlation function. Therefore, D2(r) is much more robust to both the assumed mean density and details of the selection function, making it the most reliable measure of homogeneity.
3.3 Lognormal realizations and covariance matrix
Lognormal realizations (Coles & Jones ) are an important tool for determining uncertainties in galaxy surveys. A lognormal random field is a type of non-Gaussian random field, which can be used to model the statistical properties of the galaxy distribution and simulate data sets with an input power spectrum. We have used 100 such realizations, generated using an input ΛCDM power spectrum with the fiducial parameters listed in Section . These are sampled with the survey selection function, to create 100 mock catalogues for each of the WiggleZ regions. We use these to calculate the full covariance and errors of our measurement. Jackknife resampling does not permit enough independent regions within the survey volume to give a reliable estimate of the uncertainties.





The correlation matrices for and D2(r) are shown in Figs 2 and 3, respectively, for the combined regions (see the next section) in the 0.5 < z < 0.7 redshift slice.

The correlation matrix for the measurement, obtained from lognormal realizations, for the 0.5 < z < 0.7 redshift slice.

The correlation matrix for the measurement, obtained from lognormal realizations, for the 0.5 < z < 0.7 redshift slice.
It is notable that the measurement is more correlated than D2(r). This is because
is effectively an integral of D2(r), so its covariance is effectively a cumulative sum of that of D2(r). It can also be explained by the fact that D2(r) is the logarithmic slope of
, so it does not depend on the correlations between widely separated
bins but rather the variations between neighbouring bins.
We note that the uncertainties calculated using lognormal realizations assume ΛCDM, and represent the variance we would expect to measure in a ΛCDM universe. It would not be possible to calculate uncertainties for a fractal universe, since a fractal has no defined cosmic variance as it has no defined mean density. However, calculating uncertainties this way gives a valid consistency check of ΛCDM.
3.4 Combining WiggleZ regions


We show Ccomb(i, j) for the and D2(r) measurements in the 0.5 < z < 0.7 redshift slice, in terms of the correlation matrix
, in Figs 2 and 3.
3.5 Likelihood (χ2) and parameter fitting



3.6 A new method of defining the ‘homogeneity scale' RH
As previously mentioned, the definition of a ‘homogeneity scale' is somewhat arbitrary since we expect the Universe to have a gradual approach to homogeneity. Recently, Bagla et al. () and Yadav et al. () proposed that the homogeneity scale be defined where the measured fractal dimension Dq becomes consistent with the ambient dimension within the 1σ statistical uncertainty, . They derived an approximation for Dq(r) and
, in the limit of weak clustering, for a given correlation function, and showed that both scale the same way, and so the homogeneity scale stays constant, with bias and epoch. This definition is therefore beneficial as it is not arbitrary, and is robust to the tracer galaxy population. However, in deriving
they considered only shot noise and cosmic variance from variance in the correlation function, while ignoring contributions from the survey geometry and selection function. Any real survey will have these contributions to the statistical uncertainty, which cannot be separated from the variance due to the correlation function alone. The value derived from this definition in a real survey is therefore difficult to interpret in a meaningful way, i.e. one that allows comparison with theory or with different surveys of differing volume and selection function.
We therefore introduce a different method for determining a ‘homogeneity scale' RH, which is easier to compare with theory and between surveys. Our method is to fit a smooth, model-independent polynomial to the data, and find the scale at which this intercepts a chosen value, or ‘threshold', close to homogeneity. This scale is then defined as the homogeneity scale RH. For example, RH could be the value of r at which the polynomial intercepts a line 1 per cent from or D2(r) = 3 (see Fig. 4 for an illustration). The uncertainty is found using the 100 lognormal realizations. The homogeneity scale measured this way does not depend directly on the survey errors (although the uncertainties on RH do), and is less susceptible to noise in the data, making this a preferable method that allows comparisons between different surveys. It also allows easy comparison between the data and a given model, e.g. ΛCDM, and we can check that the data converge to
or D2 = 3 as expected for a homogeneous distribution, by choosing a range of thresholds approaching homogeneity (see Fig. 9).

Homogeneity scale RH as a function of b2s8(z)2, as predicted by the ?CDM model, for different thresholds approaching homogeneity (10, 1 and 0.1?per cent from homogeneity, coloured curves from bottom to top), for (left) and D2(r) (right). The corresponding WiggleZ results are shown as error bars of corresponding colour (the errors are found using lognormal realizations). The b2s8(z)2 values of the data increase with redshift slice (so the data points from left to right are from low to high redshift). Not all redshift slices have measurements at 0.1?per cent (blue) since the data do not reach this value in those slices. The definition of homogeneity used by previous authors is where the data come within 1s of homogeneity - we show these scales as black diamonds. It is clear that this definition has much greater stochasticity than our definition, which fits a smooth curve to many data points. Indeed, this approach gives quite uninformative results in this plane and cannot be compared to the model prediction.

Illustration of our method of defining the homogeneity scale, RH, shown here for the D2(r) measurement. We first fit a model-independent polynomial (red curve) to the data (black data points). We then find where this intercepts a chosen value close to homogeneity, e.g. 1?per cent from homogeneity, D2 = 2.97 (dotted grey line). This gives us RH. We find the uncertainty in RH from the rms variance of 100 lognormal realizations (pink curves).

Scaled counts-in-spheres for the combined WiggleZ data in each of the four redshift slices (black error bars). A ?CDM model with best-fitting bias b2 is shown in blue. A fifth-degree polynomial fit to the data is shown in red. The red error bar and label show the homogeneity scale RH for the galaxy distribution, measured by the intercept of the polynomial fit with 1.01 (1?per cent away from homogeneity), with the error given by lognormal realizations. This scale is consistent with the ?CDM intercept with 1.01, labelled in blue.

Same as for Fig. but for the correlation dimension D2(r). The D2(r) measurements for the combined WiggleZ data in each of the four redshift slices are shown as black error bars. A ?CDM model with best-fitting bias b2 is shown in blue. A fifth-degree polynomial fit to the data is shown in red. The red error bar and label show the homogeneity scale RH measured by the intercept of the polynomial fit with 2.97 (1?per cent away from homogeneity), with the error given by lognormal realizations. This scale is consistent with the ?CDM intercept with 2.97, labelled in blue.

The effect of bias on a ?CDM model (left) and D2(r) model (right) at z = 0.2. Increasing the bias increases the value of on small scales, and decreases the value of D2(r) on small scales, and produces a larger homogeneity scale, as seen by the intercepts of the curves with 1?per cent of homogeneity ( and D2 = 2.97, red dotted lines).

RH for ?CDM (green) and D2(r) (purple) models at z = 0.2 with differing bias b2. Each curve corresponds to RH evaluated at a different threshold - 1, 0.1 or 0.01?per cent away from homogeneity (from bottom to top, labelled).
We can also take this further and construct a likelihood distribution for the homogeneity scale, as described in Section .
Although our choice of D2 threshold is arbitrary, by choosing the same threshold for different surveys we obtain an RH value that can be meaningfully compared, and that can be easily compared to a theoretical model. Our choice of threshold may be limited by the amount of noise in the data, however. For instance, we can measure an intercept 1 per cent away from homogeneity for the WiggleZ data, but cannot measure 0.1 per cent in two of the redshift slices, due to noise (the data do not come this close to homogeneity, although they are consistent with it within the uncertainties; see Fig. 9). None the less, we can easily choose a threshold that is possible given our data, and use this to compare with a model. The baryon acoustic oscillation (BAO) feature can also potentially affect the appropriate choice of intercept value as it causes a small distortion in D2(r) – we discuss this in Section .
For the main results in this paper, we have chosen a threshold of 1 per cent away from homogeneity, since this is about the closest threshold to homogeneity we can measure, considering the noisiness of the data.
4
CDM MODEL PREDICTION OF
AND D2(R)
In this section, we derive theoretical ΛCDM predictions for the counts-in-spheres and correlation dimension. This allows us to compare our measurements of the transition to homogeneity to the predictions of a ΛCDM model that fits WMAP data.
4.1
and D2(r)








We calculate the model D2(r) by simply applying equation to our model .
4.2 Redshift-space distortions and non-linear velocity damping


There are two forms of redshift-space distortion of relevance to our measurement, which we find are necessary for a good fit to the data.
- On large, linear scales (≳20 h−1 Mpc), the bulk infall of galaxies towards overdensities creates an enhancement in the observed power spectrum along the line of sight. This can be modelled by the linear Kaiser formula (Kaiser ):where Pδδ, L(k) is the linear matter power spectrum, f is the growth rate of structure and β = f/b is the redshift-space distortion parameter.(21)
Note that this formula assumes a perturbed FRW universe with small real-space density perturbations |δ(r)| ≪ 1.
- On quasi-linear scales (10 ≲ s ≲ 20 h−1 Mpc), the peculiar velocities resulting from the scale-dependent growth of structure distort the shape of the power spectrum, via a scale-dependent damping effect. A common way of modelling this is the ‘streaming model' (Peebles ; Fisher ; Hatton & Cole ), which combines the linear theory Kaiser formula with a velocity streaming term. We choose to use a Lorentzian term, F = [1 + (kσpμ)2]−1, for an exponential velocity probability distribution function (PDF), since Blake et al. () found this to be a good fit to the WiggleZ power spectrum for k < 0.1 h−1 Mpc. This gives us the so-called ‘dispersion model' (Peacock & Dodds ) for the full redshift-space power spectrum:Here, σp is the pairwise velocity dispersion along the line of sight. Both σp and β are parameters that must be fitted to the data. We use the values obtained by Blake et al. () in each redshift slice from fits to the WiggleZ 2D power spectrum – these are listed in Table 3.(22)
We note that the streaming model is motivated by virialized motions of particles within haloes, on much smaller scales – the so-called ‘Finger of God' effect at ≲2 h−1 Mpc. However, it is heuristic in nature and can also describe physical scales of tens of h−1 Mpc. Blake et al. () apply it this way by fitting for σp on these scales, rather than on Finger-of-God scales. We find that including it gives a significant improvement of our model fit to data.

Values of the redshift-space distortion parameter ß and the pairwise velocity dispersion sp, used in our modelling of non-linear redshift-space distortion effects. These values were obtained by Blake et al. () from fits to the WiggleZ 2D power spectrum.



Equation normally assumes a linear matter power spectrum; however, we choose to use a non-linear Pδδ(k), calculated by camb using the halofit code (Smith et al. ), since Blake et al. () found that this gave a better fit to the data.
4.3 Correcting the WiggleZ data for galaxy bias
The amplitude of the galaxy correlation function is affected by galaxy bias and redshift-space distortions, and the shape is affected by non-linear damping. Therefore, these also affect the amplitude and shape of , as well as the measured scale of homogeneity. It is possible to correct the data for bias, and hence to determine the homogeneity scale for the underlying matter distribution, by assuming a particular model, i.e. our ΛCDM+WMAP model, fitting for bias by minimizing the χ2 value, and correcting the data for this.
In our analysis, we only consider linear galaxy bias. This relates the galaxy correlation function ξg(r) to the matter correlation function ξm(r) through ξg(r) = b2ξm(r). Since we are only interested in large scales, we do not consider scale-dependent bias that may occur on small scales.


Since we assume a ΛCDM model to fit for bias, we cannot make a model-independent measurement of transition to homogeneity of the underlying matter distribution. However, it is still interesting to look at the variation of the homogeneity scale of the matter distribution with redshift, assuming ΛCDM+WMAP.
We note that our measurement of the homogeneity scale of the WiggleZ galaxies is independent of the ΛCDM modelling shown in this section. This modelling is only done so that we can show that the measurement is consistent with that expected from ΛCDM.
5 Results
5.1
and D2(r)
The measurements in each of the four redshift slices are shown in Fig. 5. The data are compared with a ΛCDM+WMAP model (described in Section ). For each successive redshift slice, the reduced χ2 values are 0.57, 0.91, 0.69 and 1.1. The first two redshift slices have 14 data bins from 12.5 to 251 h−1 Mpc, while the last two have 15 data bins from 12.5 to 316 h−1 Mpc. The data are consistent with a monotonically decreasing function, so we can fit a polynomial and find where this intercepts a chosen threshold, as per our definition of homogeneity.
The intercepts of the polynomial fit with (1 per cent away from homogeneity), which we define as the homogeneity scale RH, are shown as red error bars. The errors were determined from the 100 lognormal realizations, and correspond to the square roots of the diagonal elements of their covariance matrix. The RH values and their errors are shown in Fig. 5 and listed in Table 4, along with the values for the ΛCDM model, which are in good agreement.

Measured values of the homogeneity scale RH (where data intercept 1 per cent of the homogeneity value). The RH values shown are all for the galaxy distribution, except the values for the bias-corrected data, which are for the underlying matter distribution, i.e. b2 = 1. The bias-corrected values directly assume a ?CDM+WMAP model.
The D2(r) measurements in each of the redshift slices are shown in Fig. 6, along with a ΛCDM+WMAP model with best-fitting bias. In each redshift slice, the data range and degrees of freedom are the same as for the measurement. The reduced χ2 values in each redshift slice are 0.83, 0.90, 0.74 and 0.98. In each case, a polynomial is fitted to the data. The homogeneity scales measured where these intercept 1 per cent away from homogeneity, D2 = 2.97, are listed in Table 4 and are also in excellent agreement with the ΛCDM values.
5.2 Effect of bias and σ8(z) on RH
The homogeneity scale of the model galaxy distribution, measured at 1 per cent from homogeneity, will depend on the amplitude and shape of the correlation function, and so on galaxy bias b and σ8(z). We would also expect it to depend on redshift, since σ8(z) increases over time. These two parameters are in fact completely degenerate in the and D2(r) measurements.
Fig. 7 shows how our ΛCDM and D2(r) models vary with bias, at z = 0.2, for fixed σ8(z = 0) = 0.8. Larger bias means a larger amplitude of clustering, so that both curves are steeper on small scales. This means that the models reach homogeneity at larger radii for higher bias. This can be understood qualitatively, since highly biased galaxies are more clustered together than less biased galaxies, so we must go to larger scales before we reach a homogeneous distribution.
Fig. 8 shows how the homogeneity scale RH of the galaxy distribution varies with bias, for different intercept values approaching homogeneity, for the and D2(r) models at z = 0.2 with fixed σ8(z = 0) = 0.8. For a particular intercept value, larger bias again gives a larger homogeneity scale. Since
and D2(r) approach homogeneity asymptotically, the intercept jumps to higher values when we consider an intercept value closer to homogeneity. This image also shows the mapping between
and D2(r) homogeneity values for intercepts at 1, 0.1 and 0.001 per cent away from homogeneity. They are not identical since they are slightly different methods, but give similar results. This plot illustrates that there are many potential ways to define homogeneity, and so it is important to make consistent measurements between surveys in order for them to be comparable with each other and with theory.
Our bias-corrected and D2(r) measurements are listed in Table 4. The errors on the bias-corrected data were determined by applying a bias correction to each of the 100 lognormal realizations individually, and recalculating the covariance matrix. Since the bias correction aims to set the bias of all the realizations to b = 1, it lowers the overall variance, and so the error bars are slightly smaller than for the pure data. These measurements give our measured homogeneity scale for the matter distribution, assuming ΛCDM. We find that this scale increases with redshift, as expected in ΛCDM. However, since we are assuming ΛCDM, which has σ8(z) increasing with time, this is not a model-independent result.
As already mentioned, the effect of bias on the homogeneity scale of galaxies is degenerate with the amplitude of the correlation function, σ8(z), since the correlation function at redshift z depends on a combination of these, ξ(r, z) ∝ b2σ8(z)2. So far we have assumed a fixed value of σ8(z = 0). However, we can also make predictions independent of σ8, by finding how RH changes as a function of the combination b2σ8(z)2. This is shown in Fig. 9. We also show the WiggleZ results, which we have plotted for the best-fitting b2σ8(z)2 value in each redshift slice. This increases with redshift, so the data points from left to right go from low to high redshift. The WiggleZ results are in very good agreement with the ΛCDM+WMAP predictions. We show, for comparison, the values obtained from defining the homogeneity scale as where the data come within 1σ of homogeneity. These values have much greater stochasticity than those from our method of fitting a smooth curve to many data points, and do not give informative results in this plane.
We see that the model RH–b2σ8(z)2 curves are monotonically increasing. Since we expect σ8(z) in ΛCDM to grow over time due to growth of structure, we would therefore also expect the homogeneity scale to increase over time, for galaxies with fixed bias.
For the WiggleZ data, however, the measured homogeneity scale does not appear to decrease with redshift. This is explained by the fact that the WiggleZ galaxies have increasing bias with redshift, assuming a ΛCDM growth rate. As explained previously, this is understood to be due to the effects of Malmquist bias and downsizing on the selection of the WiggleZ galaxy population, and the colour and magnitude cuts. This counteracts the effect of decreasing σ8 with redshift. As can be seen in Fig. 9, within the b2σ8(z)2 range of the data we would not expect a significant change in RH, measured at 1 per cent from homogeneity, with redshift.
5.3 Likelihood analysis for homogeneity scale
Rather than trying to measure the scale of homogeneity directly, it can be informative to consider the likelihood that the Universe has reached homogeneity by a certain scale. We can construct a PDF for the homogeneity scale, by combining the likelihood distribution of the data with that of the definition of the homogeneity scale. This gives the probability that homogeneity is reached at a certain scale r.
We could apply a likelihood analysis to either our or D2(r) measurements. However, it would arguably be invalid for highly correlated data, such as the
measurement, since the different contributions to the probability distribution would be correlated, and we do not correct for this. The D2(r) measurement is much less correlated (see Figs 2 and 3). Also, as we have explained, D2(r) is the most robust measurement of homogeneity, so we have chosen this for our likelihood analysis.
The likelihood distribution on the data is simply given by the mean and variance of the lognormal realizations in each bin. These give the expected variance of a ΛCDM distribution sampled with the WiggleZ selection function and hence take into account both cosmic variance and shot noise. Here we assume they are Gaussian distributed by virtue of the central limit theorem. However, we also consider the true distributions provided by the lognormal realizations – this gives similar results but with larger errors (see Appendix ).

![The probability distributions pD[D2(r)] for each of the r bins in the 0.5 < z < 0.7 redshift slice. Blue-to-red gradient indicates small to large radius, from 12.5 to 316 h-1?Mpc. The area of each distribution above D2(r) = 2.97 (dotted line) gives the probability that homogeneity has been reached within 1?per cent by that value of r.](https://oup.silverchair-cdn.com/oup/backfile/Content_public/Journal/mnras/425/1/10.1111/j.1365-2966.2012.21402.x/2/m_425-1-116-fig010.jpeg?Expires=1748001943&Signature=ARav7PDaJ~rlAnx~~JYtu8yf4Oq5OUixzI8aKxAAke9vX357n~zsCu3UjPhL5PJiSK2obnYJe9N4bJcV8JjUjdaIky5ses1Kk7FelJ4XTSi1iZCy5epCfQ6NAVwiCuMJX1754tz~txX6z~Z3QasFHL-XXdaNtqFX7ZnvlZZWOAVuEbdiT8suvlbrQUzLOf0o~-tkulDqBkJnp-7wK00gBy7OKSq8iDF~oRSFzSfYymGWpmXrVUBLMlsTRkBncQ7FQpunvm3m~kJMaA54EpYOTqS10hK~-XXeXrPHYzQ04Zi8s9mTJfW8u2YCCXsTlhVARdGDGEv-g0IptbquUbF38Q__&Key-Pair-Id=APKAIE5G5CRDK6RD3PGA)
The probability distributions pD[D2(r)] for each of the r bins in the 0.5 < z < 0.7 redshift slice. Blue-to-red gradient indicates small to large radius, from 12.5 to 316 h-1?Mpc. The area of each distribution above D2(r) = 2.97 (dotted line) gives the probability that homogeneity has been reached within 1?per cent by that value of r.
We also expect there to be a likelihood distribution on the D2(r) value we would measure for a perfectly homogeneous distribution, due to cosmic variance and shot noise caused by our selection function. We can represent this by a likelihood distribution on the homogeneity scale – this would not be a simple delta function at D2(r) = 3, but would have some spread. This distribution should be one-sided – i.e. we do not expect to measure D2(r) > 3, only D2(r) ≤ 3, if we have a distribution that approaches homogeneity. We might expect it to be represented by the variance in the random catalogues, which are essentially homogeneous distributions sampled with the WiggleZ selection function. However, the same sources of noise are also present in the lognormal realizations, so we would potentially double count errors if we also used lognormal realizations to determine the likelihood distribution of the data. This means we cannot easily determine the true variance in the value of D2(r) we would expect to measure for a homogeneous distribution, independently of the variance of the data.



The PDFs for the homogeneity scale, for WiggleZ galaxies in each redshift slice, are shown in Fig. 11. We have interpolated between the data bins in order to obtain smoother PDFs. We find the most probable RH values from the mean of the distributions. These are all between 70 and 81 h−1 Mpc, and are listed in Table 4. They represent the most probable scale at which the galaxy distribution reaches 1 per cent of homogeneity. We also list the values found for the bias-corrected data, which give the most probable homogeneity scales for the matter distribution, assuming ΛCDM+WMAP.

Probability distributions for the scale of homogeneity, p(RH), for WiggleZ galaxies in each of the four redshift slices. The homogeneity scale is defined as the scale where the data reach values 1?per cent away from D2 = 3, i.e. D2 = 2.97.
6 ROBUSTNESS OF HOMOGENEITY MEASUREMENT
In this section, we address several issues that could potentially influence our measurement of the homogeneity scale, and perform tests to ensure the robustness of our results. We base our tests on the 15-hr 0.5 < z < 0.7 region, which is the largest and most populated WiggleZ subregion, but the results are applicable to the entire survey.
6.1 Fractal model test of selection function and boundary effects
A major potential source of bias in our results is the method used to correct for edge effects and the selection function of the survey. We have used 100 random catalogues to correct each individual WiggleZ measurement, as described in Section . However, this method can potentially bias homogeneity measurements, since it weights measurements by the volume of spheres of radius r included in the survey, and so assumes a homogeneous distribution outside the survey (e.g. Coleman & Pietronero ; Sylos Labini et al. ). It is therefore important to check that this is not imposing any distortion in our measured correlation dimension, so producing a ‘false relaxation' to homogeneity.
To test this, we apply our correction method to a range of fractal distributions of known correlation dimension. This has been done previously by a number of works, e.g. Lemson & Sanders (), Provenzale, Guzzo & Murante () and Pan & Coles (). This allows us to check that our method returns the correct input correlation dimension up to the largest scales we measure in the survey, and to quantify any distortion that may occur.

Our procedure for using β-models to test our analysis method is as follows.
We choose a range of D2 values (2.7, 2.8, 2.9, 2.95 and 2.97), and for each we generate 100 fractal galaxy distributions, with a boxsize of (L0h−1 Gpc)3, where L0 is the length of the longest side of the WiggleZ 15-hr 0.5 < z < 0.7 selection function grid (623.5 h−1 Mpc).
We then sample each distribution with the WiggleZ selection function for the 15-hr 0.5 < z < 0.7 region. We normalize the resulting distribution to give the same number of points as WiggleZ galaxies in this region. This gives us fractal mock catalogues, and we then measure D2 for these in the same way as for the WiggleZ data, correcting the counts-in-spheres measurements with random catalogues. This gives a result that is influenced by both the WiggleZ selection function, and our correction method.
Fig. 12 shows the mean and D2(r) results for the different fractal distributions up to D2 = 2.95, with the WiggleZ selection function and correction method applied. Even up to D2 = 2.95, they are clearly distinguishable from ΛCDM and the WiggleZ data. The D2 values measured are consistent with the input values up to at least 200 h−1 Mpc, well above the homogeneity scales measured for the data. These results indicate that the WiggleZ selection function and correction method do not have a significant effect on the measured correlation dimension up to the scales we measure for homogeneity.
![Fractal model comparisons of the measured correlation dimension from the WiggleZ 15-hr region, 0.5 < z < 0.7 redshift slice. The WiggleZ data (black error bars) and ?CDM model (blue curve) are compared with several different ß-models with different fractal dimension (D2 = 2.7, 2.8, 2.9 and 2.95), which have been sampled with the 15-hr selection function and analysed in the same way as the WiggleZ data (coloured error bars). The uncertainties shown are the error in the mean of 100 fractal realizations. The input fractal dimensions are shown as dotted lines with corresponding colours. The best-fitting D2(r) value fits over the range r = [80, 300]?h-1 Mpc are shown as solid lines.](https://oup.silverchair-cdn.com/oup/backfile/Content_public/Journal/mnras/425/1/10.1111/j.1365-2966.2012.21402.x/2/m_425-1-116-fig012.jpeg?Expires=1748001943&Signature=uk5TG5C46FPgsAe3zmPHT4RopodDXvzHlHaXridW3-qgGxZRPFv2twa888DMYpxxfR0~YtGsTT19g-g0bL6Vv14i5d-LLLTjK9Hekbm3jq-gkyrDcfysTKhvyGNhRbzHhMAejetgcc7-~Fa9~CC5woRzks1vM1A7g-xi13DNVTfiuPth3EK6c2xI9a~Jg39d8PGC4ebGS3VleACrnFP6q23bs4JB7ZKzTMlwpROZexet2GVCuCPURd0zka7hQic0TyX5OsRckzSp9Zhrpq6rU81dIYtQaQbfi5oBz2DLM~0M6s~VKhx5ZVMHWk0MEoM5zucs2-WXcUuFPZ2ouVY7XQ__&Key-Pair-Id=APKAIE5G5CRDK6RD3PGA)
Fractal model comparisons of the measured correlation dimension from the WiggleZ 15-hr region, 0.5 < z < 0.7 redshift slice. The WiggleZ data (black error bars) and ?CDM model (blue curve) are compared with several different ß-models with different fractal dimension (D2 = 2.7, 2.8, 2.9 and 2.95), which have been sampled with the 15-hr selection function and analysed in the same way as the WiggleZ data (coloured error bars). The uncertainties shown are the error in the mean of 100 fractal realizations. The input fractal dimensions are shown as dotted lines with corresponding colours. The best-fitting D2(r) value fits over the range r = [80, 300]?h-1 Mpc are shown as solid lines.
It is notable that the size of the error bars gets smaller for models with larger D2. This is a real effect in the model; for larger D2, more boxes survive at each iteration, so there are a smaller number of possible configurations for the final distribution, resulting in lower variance for a box of a particular volume.
To quantify how well we can exclude fractal models, we fit a line of constant D2 to each set of fractal data, over the range [80, 300] h−1 Mpc (shown in Fig. 12). This gives us the best-fitting D2 value we would expect to measure for each fractal distribution over this range, taking into account bias from the selection function. We then find the formal probability of these values fitting the WiggleZ data. Doing this, we find we can exclude a fractal dimension of D2(r) = [2.9, 2.95, 2.97] at the [19,6,4]σ level. In other words, we can exclude fractal distributions with dimension D2(r) < 2.97 at over 99.99 per cent confidence on scales from 80 to 300h−1 Mpc.
Our results agree with those of Lemson & Sanders (), Provenzale et al. () and Pan & Coles (), who also find that for samples on scales larger than the homogeneity scale, boundary corrections do not have a significant effect on the analysis. A further check would be to test different types of fractal model other than the β-model, but we leave this for future work. We also note that we still assume an FRW metric in our fractal analysis; an improved consistency check would be to calculate the actual metric in these fractal models (see Section ), but this is beyond the scope of this paper.
6.2 Comparison with the GiggleZ N-body simulation
We have also compared our results with a ΛCDM cosmological N-body simulation. This allows us to check that our analytical ΛCDM+WMAP model (incorporating Kaiser and streaming models for redshift-space distortions) is consistent, over the relevant scales, with a full simulation including non-linear effects. It also provides a further test of selection function effects, since we can show that homogeneity measurements of a ΛCDM distribution are not distorted when the WiggleZ selection function is applied.
The GiggleZ (Giga-parsec WiggleZ) simulation (Poole et al., in preparation) is a suite of dark matter N-body simulations run at Swinburne University of Technology, designed for theoretical analyses of the WiggleZ data set. It was run using a modified version of the N-body code gadget-2 (Springel, Yoshida & White ) using a WMAP5 cosmology. We use the main simulation, which has a volume of (1000 h−1 Mpc)3 and 21603 particles of mass 7.5 × 109h−1 M⊙.
Halo finding for GiggleZ was performed using subfind (Springel et al. ), which utilizes a friends-of-friends (FoF) algorithm to identify coherent overdensities of particles and a substructure analysis to determine bound overdensities within each FoF halo. For our analysis, we rank-order the resulting subfind substructure catalogues by their maximum circular velocity (Vmax) and select a contiguous subset of 250 000 haloes (selected to yield a number density comparable to the survey) over a range of Vmax chosen to yield a bias comparable to that of WiggleZ. We use a catalogue at a redshift of z = 0.593, corresponding to the mid-redshift of the WiggleZ 0.5 < z < 0.7 redshift slice.



We then calculate and D2(r) using the following two different methods.
Using the full GiggleZ box. We correct the measurement using a random distribution within the same volume box, with 100 times the number of galaxies as the GiggleZ sample.
Applying the WiggleZ 15-hr 0.5 < z < 0.7 selection function to GiggleZ. This creates a mock WiggleZ survey containing 10 830 galaxies. We correct the measurement using the random catalogues used for the WiggleZ data. This allows us to see the effects induced purely by the selection function.
The results are shown in Fig. 13. The full GiggleZ data set is very consistent with both the WiggleZ data and the ΛCDM model. The consistency with the model indicates that the implementation of redshift-space distortions in the model, described in Section , has a good level of accuracy, to scales as small as ∼20 h−1 Mpc. The deviation of the smallest scale bin is a resolution effect, since the GiggleZ catalogue is sparser than WiggleZ (with 10 830 galaxies, versus 17 928 WiggleZ galaxies in the same region and redshift slice). The GiggleZ results both with and without the WiggleZ selection function are also consistent within the size of the WiggleZ error bars, showing that the selection function and correction method do not have a significant effect. It can be seen that the selection function causes a few data points in the plot to go below 1, and in the D2(r) plot to go above 3. This shows that adding shot noise can produce this apparent unphysical effect, explaining why this is also seen in some of the WiggleZ results (Figs 5 and 6).

Comparison of the GiggleZ N-body simulation with WiggleZ, for the 15-hr region 0.5 < z < 0.7 redshift slice. The results are shown on the left, and D2(r) on the right. The WiggleZ data are shown as black data points, and a ?CDM model is shown in blue. The results for the full GiggleZ box are shown as the red crosses. The green crosses show the results for the GiggleZ simulation sampled with the WiggleZ 15-hr 0.5 < z < 0.7 selection function. The measured homogeneity scale RH is indicated for each.
6.3 Comparison of different correction methods
To further demonstrate the robustness of our correction method, we illustrate the results obtained for two alternative correction methods – first, using only complete spheres for the counts-in-spheres measurements (the ‘exclusion' method mentioned in Section ), but still correcting for the selection function, and secondly, using no correction for the selection function. By using the exclusion method, we do not have to deal with survey edge effects; however, since WiggleZ is not volume limited, it is still necessary to use random catalogues to correct for the selection function. That is, we recalculate equation but with G equal to the number of WiggleZ galaxies at the centre of complete spheres of radius r. We show this result in Fig. 14, for the 15-hr 0.5 < z < 0.7 region, as red error bars, where the uncertainty is calculated using our lognormal realizations. We compare this to the result using our correction method (black error bars). The red error bars are consistent with the black error bars, but show more scatter and have higher noise, which increases for larger radius, reflecting the lower statistics where fewer spheres contribute to the measurement. The measurements must also be cut off at a lower radius, since not enough larger spheres fit inside the survey region.

Illustration of the result using different correction methods, for the 15-hr 0.5 < z < 0.7 region. Black error bars show the result using our correction method with random catalogues. The red error bars are obtained when only complete spheres are used in the counts-in-spheres measurement while still correcting for the selection function. The, purple error bars show the result for complete spheres, with no selection function correction. In each case, the uncertainties are calculated using lognormal realizations. The red and purple error bars are shifted slightly to the right for clarity.





7 Discussion
Our WiggleZ and D2(r) results show a very strong agreement with an FRW-based ΛCDM model. However, one of the strongest arguments against previous homogeneity measurements is that the method of correcting for survey selection functions, such as using random catalogues as we do, can distort the data, producing a ‘false relaxation' to homogeneity. We have therefore tested this by applying the WiggleZ selection function to a range of inhomogeneous fractal models and a ΛCDM N-body simulation. We have shown that there is no significant impact on our homogeneity measurement, up to at least 200 h−1 Mpc. In addition, we have compared the results from our correction method to an analysis using only complete spheres, with and without correcting for the selection function, and have shown that these are consistent, further demonstrating the robustness of our result. We can rule out fractals with fractal dimension up to D2 = 2.97, on scales between 80 and 300 h−1 Mpc, at over 99.99 per cent confidence.
We can also be confident that our result is robust to any assumptions in modelling the WiggleZ survey selection function, since Blake et al. () showed that even extreme variations in modelling the angular completeness produce only ∼0.5σ shifts in estimates of the power spectrum. Changes in the parametrization of the redshift distribution were shown to cause larger deviations, but only at scales >200 h−1 Mpc, which are well above the scales on which we measure homogeneity.
Our result is a very good consistency check of ΛCDM. However, it is not independent of the assumption of the FRW metric. A complication for all homogeneity measurements is that we can only observe galaxies on our light cone, and not on spatial surfaces. Maartens () points out that it is therefore not possible to make a homogeneity measurement without making some assumptions, such as the FRW metric (to convert redshifts to distances) and the cosmological principle. Indeed, we must always assume an FRW-based model, ΛCDM in our case, to convert redshifts and angles to distance coordinates. This is also problematic considering that inhomogeneities produce perturbations in the FRW metric, so can potentially distort distance measurements by affecting the paths travelled by light rays (e.g. Wiltshire ; Meures & Bruni ). However, it seems highly unlikely that our measurements would so closely agree with FRW-based ΛCDM [in both the amplitude and shape of the and D2(r) curves], if the distribution were, actually, inhomogeneous up to the largest scales probed, and we had incorrectly assumed an FRW metric. We have tested our method using a fractal model, and shown that this gives a completely different form of these curves. Our assumption of FRW also seems reasonable, since we know that distance measurements from Type Ia supernovae fit the Hubble diagram well to first order up to z ∼ 1.4 (Conley et al. ; Davis et al. ), so that any perturbations due to inhomogeneities can only be a second- or higher-order effect. All this means that we can take our results to be a strong consistency check of FRW-based ΛCDM.
It would be possible to test this further by making isotropy measurements in thin redshift shells, over a range of redshifts, without converting them to distances. We could also directly test the assumption of FRW, by calculating the non-FRW metrics of our fractal models and testing our analysis on these. We could also calculate the effects of metric perturbations for ΛCDM, including the effects of backreaction, as in e.g. Buchert (), Li & Schwarz (), Wiltshire () and Behrend, Brown & Robbers (). This would allow us to test the effect of incorrectly assuming an FRW metric. Alternatively, several possible consistency tests of homogeneity and the Copernican Principle that do not assume FRW have been suggested by e.g., Clarkson et al. (2008); Shafieloo & Clarkson (); Maartens (); Heavens et al. (). However, we leave all these suggestions for future work.
An alternative way of defining the ‘homogeneity scale' was recently suggested by Bagla et al. () and Yadav et al. (), as where the measured fractal dimension Dq becomes consistent with the ambient dimension within the 1σ statistical uncertainty, . They make predictions for a ΛCDM model, by deriving an approximation for Dq and
, given a particular correlation function, and showing how these scale with sphere size. Using this, they predict that the ‘true' homogeneity scale in ΛCDM is 260 h−1 Mpc. They also predict that Dq and
scale the same way with the correlation function, and so their definition of homogeneity does not change with bias or redshift. Their definition is therefore beneficial, as it is robust to the tracer galaxy population. It is also not arbitrary, and indeed the scale above which the fractal dimension is consistent with homogeneity within cosmic variance is arguably a physically meaningful scale to define as homogeneous, since above this scale the distribution cannot be distinguished from a homogeneous one. However, as we have pointed out (Section ), it is difficult to apply their definition to a real measurement, since their approximation for
only accounts for the variance of the correlation function (in the limit of weak clustering) and shot noise, but ignores errors due to survey geometry and the selection function. These additional contributions mean that real errors will always be larger, and so will always measure a smaller homogeneity scale than the ‘true' one. Since these errors will be different for different surveys, and cannot be separated out from the variance in the underlying correlation function, homogeneity measurements made in this way cannot be easily compared between different surveys, or with theory. The benefit of our method of defining the homogeneity scale, even if we have to make an arbitrary choice about the value of ΔDq we accept for homogeneity, is that it can be used to easily compare the results from different surveys and with a theoretical model.
A homogeneity scale below 100 h−1 Mpc may also seem to contradict the fact that the correlation function has a known feature, the BAO peak at ∼105 h−1 Mpc (Eisenstein et al. ; Percival et al. ; Beutler et al. ; Blake et al. ). However, the BAO peak has only a small impact on the counts-in-spheres statistic. It is visible in our ΛCDM prediction for D2(r) [which is more sensitive to it than , since it is a differential measurement], as a small dip at just over 100 h−1 Mpc (see Fig. 7). It means that the D2(r) curve does not increase monotonically around this scale, so we must be careful if attempting to measure an intercept with homogeneity that lies close to this. The magnitude of the distortion due to the BAO peak, ΔD2, is of order ∼0.01 for b ∼ 1 so does not affect our homogeneity scale measured at D2 = 2.97. However, it is more significant for more highly biased tracers (as also pointed out by Bagla et al. ), and for a highly biased population, such as LRGs, it may be necessary to measure the homogeneity scale at a D2 value closer to 3, to avoid this region.
It has been pointed out (e.g. Sylos Labini & Pietronero ) that measurements of large structures in galaxy surveys are seemingly at odds with a homogeneity scale below 100 h−1 Mpc. Previous surveys have detected structures on scales much larger than this (de Lapparent, Geller & Huchra ; Geller & Huchra ). The largest observed structure in the Universe, the Sloan Great Wall, which is 320 h−1 Mpc long (Gott et al. ), appears inconsistent with the existence of a homogeneity scale below 100 h−1 Mpc. However, this is not incompatible with our results, since we show the scale where the data have D2 ≥ 2.97, and it is not impossible to have fluctuations on scales larger than this. Also, these large structures, including the Sloan Great Wall, are usually filamentary, whereas we have measured a volume statistic which averages over fluctuations.
8 Conclusion
We have measured the large-scale transition to homogeneity in the distribution of galaxies, using the WiggleZ survey, in four redshift bins between z = 0.1 and 0.9. We measured the mean, scaled counts-in-spheres and the correlation dimension, D2(r), and found these to be in excellent agreement with a ΛCDM model with WMAP parameters, including redshift-space distortions. We also presented a new, model-independent method for determining the homogeneity scale RH from data. This involves fitting a polynomial curve to the data, and finding where this intercepts chosen values close to homogeneity. This is a more reliable method than finding where the data come within 1σ of homogeneity, since it does not depend directly on the size of the error bars and is less susceptible to noise. It also allows a direct comparison between data and theory, and between different surveys of differing bias and redshift.
We summarize our results as follows.
Our
and D2(r) results show a very strong agreement with an FRW-based ΛCDM+WMAP model incorporating redshift-space distortions. They show a clear transition from an inhomogeneous, clustered distribution on small scales, to a homogeneous one on large scales. This transition matches that of the ΛCDM model. We have thereby conducted a very stringent consistency check of ΛCDM.
If we define the ‘homogeneity scale' RH as the scale where the data become consistent with homogeneity within 1 per cent, then from a likelihood analysis of D2(r), we measure RH to be 71 ± 8 h−1Mpc at z ∼ 0.2, 70 ± 5 h−1Mpc at z ∼ 0.4, 81 ± 5 h−1Mpc at z ∼ 0.6 and 75 ± 4 h−1Mpc at z ∼ 0.8. These values are consistent with those of the ΛCDM+WMAP model with best-fitting bias, of RH =76, 78, 81 and 78 h−1 Mpc.
We find that the homogeneity scale of our ΛCDM+WMAP model increases with clustering amplitude b(z)σ8(z). For a population with fixed bias, we therefore predict the homogeneity scale to grow over time, since σ8(z) increases due to growth of structure. The bias of the WiggleZ galaxies increases with redshift, so the measured RH values do not change with redshift. If we correct our data for bias, assuming ΛCDM, then we measure a homogeneity scale for the matter distribution that increases over time, consistent with our ΛCDM+WMAP model.
The WiggleZ results are in excellent agreement with those of the GiggleZ N-body simulation incorporating redshift-space distortions. It is also in excellent agreement with our analytic ΛCDM+WMAP model, showing that our model for redshift-space distortions is accurate down to scales as small as 20 h−1 Mpc.
We can exclude a fractal with fractal dimension up to D2 = 2.97, on scales between ∼80h−1 Mpc and the largest scales probed by our measurement, ∼300 h−1 Mpc, at 99.99 per cent confidence.
By applying our analysis to the GiggleZ simulation, as well as a suite of fractal distributions of differing fractal dimension, we have shown that our result is not significantly distorted by the WiggleZ selection function and our method of correcting for it. We also show that we obtain consistent results even using different correction methods, i.e. using only complete spheres for the measurement, with and without correcting for incompleteness. This therefore confirms the reliability and robustness of our results.
Acknowledgments
MIS would like to thank Florian Beutler for many helpful comments on the paper. We also thank Eyal Kazin, David Parkinson, Luigi Guzzo and Andy Taylor for helpful discussions. MIS acknowledges support from a Jean Rogerson Scholarship and a UWA Top-up Scholarship from the University of Western Australia. GBP acknowledges support from two Australian Research Council (ARC) Discovery Projects (DP0772084 and DP1093738). The Centre for All-sky Astrophysics (CAASTRO) is an Australian Research Council Centre of Excellence, funded by grant CE11E0090.
GALEX is a NASA Small Explorer, launched in 2003 April. We gratefully acknowledge NASA's support for construction, operation and science analysis for the GALEX mission, developed in cooperation with the Centre National d'Etudes Spatiales of France and the Korean Ministry of Science and Technology.
The WiggleZ survey would not have been possible without the dedicated work of the staff of the Australian Astronomical Observatory in the development and support of the AAOmega spectrograph, and the running of the AAT.
References
A Appendix
GAUSSIAN VERSUS TRUE PROBABILITY DISTRIBUTIONS pD[D2(r)]
In our likelihood analysis for the homogeneity scale (Section ), we assume that the distribution of the 100 lognormal realizations in each bin, pD[D2(r)], is Gaussian. This allows us to interpolate between the data points and errors, thereby creating more finely spaced pD[D2(r)] distributions, in order to determine a smoother PDF for the homogeneity scale, P(RH ≤ r). However, it is not obvious that these distributions should be Gaussian. We therefore repeat the analysis, but use the true distributions given by the lognormal realizations.
This gives the pD[D2(r)] distributions shown in Fig. 15. They are not smooth Gaussians, although they are close to Gaussian. Their resolution is limited by the number of lognormal realizations, so they could be improved by using more lognormal realizations, although we do not do this here.
![The probability distributions pD[D2(r)] for each of the r bins (blue-to-red gradient indicates small to large radius) in the 0.5 < z < 0.7 redshift slice. Unlike in Fig. , we do not assume these are Gaussian distributions; rather, we plot the distributions given by our 100 lognormal realizations.](https://oup.silverchair-cdn.com/oup/backfile/Content_public/Journal/mnras/425/1/10.1111/j.1365-2966.2012.21402.x/2/m_425-1-116-fig015.jpeg?Expires=1748001943&Signature=XMUrjwCIbWou5KX0z3HulivhSeuSjvOUQ9zixmLLn~HGkwRVrHlUyDL-~BXNy5LjtWZCWmeLpoE~ZSEycNgd~ubNMprYanlE3xuEx5pL61rqI0VisiAZNNRehvQNqCpRf69bCasH1YLfUU1pkXiFJ~EXZEwjDFPxOnlv0LY~5axXsS4132TdQJt5nZ5ZZnDI-JkeMk~B8tmjRnZE7VbNo1CGI1GmRac74~mAW5DgEnu0EIQfmCCuaR7VXyxrj-q0lP9CRSDT9ovP-OKeA-8yvqmvoY4UlwMWMJ5zcXID1h4BM3S9mTZD2DiXebFjc20xiZgzfu7khG7~~cFDDobCVA__&Key-Pair-Id=APKAIE5G5CRDK6RD3PGA)
The probability distributions pD[D2(r)] for each of the r bins (blue-to-red gradient indicates small to large radius) in the 0.5 < z < 0.7 redshift slice. Unlike in Fig. , we do not assume these are Gaussian distributions; rather, we plot the distributions given by our 100 lognormal realizations.
We then use these to calculate the PDF for the homogeneity scale, P(RH ≤ r), in the same way as we did in Section for the Gaussian distributions. This gives the PDFs shown in Fig. 16, for each redshift slice. The mean values and errors are shown in Table 5, along with those from our original analysis. The values are very similar, though the errors are larger. This is because we cannot interpolate between data points as easily, and there are a finite number of lognormal realizations contributing to the distribution for each data point. This means the distribution is effectively smoothed, giving larger uncertainties.
![Probability distributions for the homogeneity scale, p(RH), for WiggleZ galaxies in each of the four redshift slices. These are calculated from the probability distributions pD[D2(r)] of the 100 lognormal realizations, rather than assuming Gaussians.](https://oup.silverchair-cdn.com/oup/backfile/Content_public/Journal/mnras/425/1/10.1111/j.1365-2966.2012.21402.x/2/m_425-1-116-fig016.jpeg?Expires=1748001943&Signature=Ks9v2kI88VBOIcNavJ3GiIRntMvgPEMT14EH6gjMf0b1E6QdAnLOwCkpYjcIfE81fr0XDZ8HvHXvXTFuyfzbh-wrOfqLfNUxyXP3mYF3~iv~MbuATWM0EUecwhDjIYJQerxSahbtiS6MQONz1NfgwE8BjGJNAeBwj4bhP04BMnKZiQpZDEsSpCv6VNNxAaknF0WB5DLJ-HMfA1NBiOBgrvChismNXSoDxrjLnU6~gau3U~1yrK1F2OpQNYzlneggA70EDppfGaAkw0ge1JRXUANUpLGGqV17eqmg8Y49bnDIIL324o9P5DHOE9rYCpmJlIHvBRNKam2vLBxu546RPg__&Key-Pair-Id=APKAIE5G5CRDK6RD3PGA)
Probability distributions for the homogeneity scale, p(RH), for WiggleZ galaxies in each of the four redshift slices. These are calculated from the probability distributions pD[D2(r)] of the 100 lognormal realizations, rather than assuming Gaussians.
![Comparison of the most probable RH values from a likelihood analysis using the true pD[D2(r)] distributions from lognormal realizations and assuming Gaussian distributions.](https://oup.silverchair-cdn.com/oup/backfile/Content_public/Journal/mnras/425/1/10.1111/j.1365-2966.2012.21402.x/2/m_425-1-116-tbl005.jpeg?Expires=1748001943&Signature=WABGujf4vV2aQiPevjEW02iMv47m1LJi9owhubDMyvWUE4eUoAmbTDN259ffoia2JUbYXihdPv3R0IAIt2gj9ZhHB3RsXRQ7VsV363ts0u3r3j6d8BC1-8xFZ6gMW~ulf2d9csysOObfDpUPsGIykhMd3pM26gcWEf~mS38yeAYPqByCJgERjyPZCvhyAtzNtyrkscsEMgNTTIgmpW2L-m-~EABZDaoyqN004ZWRBI4Dna1EvrseaOJr0oOM5O8~4zxLpkfuBxwe9dHkCDtRyWvBOwsQLU38dMDbxhjb6Xmh7FXBsM1uz04zeymqky1FQqTSSQNKz1IHhNlHSv1g3g__&Key-Pair-Id=APKAIE5G5CRDK6RD3PGA)
Comparison of the most probable RH values from a likelihood analysis using the true pD[D2(r)] distributions from lognormal realizations and assuming Gaussian distributions.
Author notes
In some cases, the WiggleZ data do fall below or above D2 = 3; this can be explained as the effect of shot noise introduced by the selection function rather than a physical effect, as shown in our comparison with the GiggleZ simulation in Section .