The paper provides end-to-end DESI DR1 LSS catalog construction: target selection, redshift success cuts, veto masks, random catalogs, and systematic weights.
Briefing
This DESI Collaboration paper addresses a practical but central question for precision cosmology: how to construct and validate large-scale structure (LSS) galaxy and quasar catalogs from DESI Data Release 1 (DR1) such that the measured two-point clustering statistics (correlation functions and power spectra) can be reliably compared to theoretical models. The question matters because raw observed number densities are modulated by many non-cosmological effects—survey geometry, fiber assignment incompleteness, imaging systematics, and spectroscopic success variations—each of which can imprint spurious correlations or bias the inferred cosmological parameters if not modeled correctly.
The paper’s contribution is methodological and infrastructural. It defines the DESI DR1 target samples (BGS, LRG, ELG, QSO), specifies the spectroscopic redshift selection criteria used to define “good” tracers, constructs the corresponding random catalogs that encode the survey footprint, and develops a weighting and correction framework to remove or marginalize known observational systematics. It also describes the measurement pipelines for two-point statistics in both configuration space and Fourier space, including the role of window functions, normalization constraints, and integral-constraint effects. Finally, it validates the resulting “raw” two-point measurements by comparing DESI DR1 data to mock catalogs that include realistic DESI fiber assignment.
Methodologically, the study is a large-scale survey data processing and validation effort rather than a controlled experiment. The data source is DESI DR1, using observations through June 14, 2022. The sample construction begins with photometric target selection from Legacy Surveys DR9 imaging, split into North (BASS/MzLS) and South (DECam/DECaLS) regions, with WISE W1/W2 used across the sky. The paper then adds spectroscopic information from the DESI “iron” reduction pipeline (Redrock redshifts, plus QSO classifiers such as QuasarNET and an MgII afterburner). A special reprocessing substitution is applied for data taken on Dec. 12, 2021 due to a calibration bug, affecting 8 dark and 9 bright tiles.
The final clustering samples are defined by tracer-specific spectroscopic success criteria and redshift cuts. For BGS, success requires ZWARN = 0 and , with redshift selection and an absolute magnitude cut (using Fastspecfit-based k-corrections and an evolution correction). This reduces the number of successful redshifts from 4,036,190 to 485,331 before the redshift cut, and then to 300,043 used in the final analysis. For LRG, success is ZWARN = 0 and , with split into three bins. For ELG, success is and split into two bins. For QSO, success is “not rejected by the quasar catalog,” using Redrock/MgII/QuasarNET identifications, with a broad selection but primary clustering using .
The paper reports final sample sizes and key completeness metrics (Table 2). After vetoes and selection, the number of good redshifts are: BGS 300,043; LRG 2,138,627; ELG 2,432,072; and QSO 1,223,391 (with the primary QSO subset 856,831 in ). Spectroscopic success rates within the footprint are high for BGS (98.9%) and LRG (99.1%), moderate for ELG (72.7%), and lower for QSO (66.8%).
A major methodological component is the construction of random catalogs and the correction of selection-function variations. Randoms are generated at uniform angular density (2500 deg per random file) over the DESI footprint defined by “reachable” targets under good hardware conditions. The paper applies multiple veto masks: hardware vetoes (bad fibers, low template signal-to-noise thresholds TSNR2, and instrument flags), a priority veto (e.g., QSO and rare strong-lens candidates remove area from lower-priority samples), and imaging vetoes (bright object masks and Healpix-based cuts on imaging property tails). The paper quantifies footprint losses; for example, hardware veto removes 3.1% of the dark-time footprint and 2.3% of the bright-time footprint. Priority veto removes 1666.7 deg (20.3%) for LRG & ELG and 38.5 deg (0.5%) for QSO.
Fiber assignment incompleteness is treated by defining assignment completeness and decomposing it into weights based on tile/fiber competition. The fiducial completeness weight is (Eq. ), with additional handling of tile-group effects via applied to randoms. The paper emphasizes that this fiducial approach is not strictly unbiased at small angular scales because of physical fiber collision constraints; therefore, DESI’s default two-point analysis removes pairs with angular separation (“-cut”) and incorporates the resulting bias through window matrices.
Imaging systematics are corrected using tracer-dependent regression methods. BGS uses linear regression with three maps (r-band depth, stellar density, HI column density). LRG uses linear regression with five maps in BASS/MzLS and one additional map in DECam regions depending on redshift bin. QSO uses a random-forest regression (Regressis). ELG uses a SYSNet neural-network regression. The paper validates these choices with null tests: it computes statistics of normalized projected density versus imaging-property bins, comparing data to uncontaminated mocks. It reports large improvements in the null after applying weights (e.g., for BGS, improvement close to a factor of 2 in DECam; for QSO, overall improvement greater than a factor of five in DECam). It also identifies residual concerns: for LRG, some residual trends in DECam are driven by CIB contamination components of , and for ELG the imaging systematic impact is the most severe.
Spectroscopic systematics are handled via redshift-failure weights , modeled as a function of TSNR2 (and for some tracers, fiberflux and redshift). The paper reports that despite significant trends in success rates with observing conditions, the impact on two-point clustering is negligible for DR1 when using these weights.
Two-point statistics are measured using Landy–Szalay estimators for correlation functions and FKP-based estimators for power spectra, with multipoles . The paper describes the use of multiple random catalogs to reduce noise, the role of -cut, and the computation of window matrices that map theoretical predictions to binned measurements. For power spectra, it also details corrections for radial integral constraint (RIC) and angular integral constraint (AIC) induced by imaging weights, using polynomial templates fit to differences measured in EZmocks and AbacusSummit cut-sky mocks.
The key validation results compare DESI DR1 “raw” two-point measurements to the mean of 25 “altmtl” mocks (mocks with realistic fiber assignment and the same LSS pipeline). The paper’s headline conclusion is that configuration- and Fourier-space two-point measurements are generally consistent to within about a 2% factor in the inferred real-space overdensity field. In the detailed comparisons (Table 9), BGS shows excellent configuration-space agreement (e.g., for over ), while Fourier-space agreement is somewhat worse but improves when restricting to . LRG configuration-space agreement is generally strong; the largest values occur in Fourier space for the highest redshift bin, but improve when restricting to (e.g., LRG3 monopole under with a bias scaling of 0.99). ELG exhibits the most notable tension: configuration-space for in has PTE , and Fourier-space mismatches are largest at low , consistent with residual imaging systematics. QSO mismatches are strongly scale dependent in Fourier space, with improved agreement when restricting to (e.g., QSO monopole for without scaling, improving to with a scaling factor 0.9882).
Limitations are inherent to the methodology. The paper’s validation uses a finite number of mocks (25 altmtl mocks for comparisons; 1000 EZmocks for covariance validation), and the mocks only approximately reproduce all aspects of the real data (notably fiber assignment realism and imaging systematics). It also acknowledges that the fiducial completeness weighting is not unbiased at small scales and therefore relies on the -cut and window-matrix modeling. For small-scale clustering, the paper points to an alternative catalog version (v1.5pip) using PIP (pairwise inverse probability) weights and angular up-weighting, which is computationally more demanding and noisier on large scales.
Practically, the results matter for anyone performing DESI DR1 clustering analyses—BAO, full-shape RSD, and primordial non-Gaussianity studies—because they provide the public LSS catalogs, the recommended weights, and the measurement/correction pipeline needed to avoid systematic biases. The paper also provides guidance on robustness testing: analyses should assess sensitivity to imaging systematic weights (especially for ELG), spectroscopic success modeling, and residual integral-constraint effects. The intended audience includes DESI collaboration analysts and external researchers using DESI DR1 LSS products, as well as method developers interested in survey systematics modeling.
Overall, the paper’s core contribution is an end-to-end, validated framework for turning DESI DR1 observations into clustering-ready catalogs and two-point measurements, with explicit quantification of sample sizes, completeness, vetoed footprint fractions, weighting schemes, and the degree of agreement between data and realistic simulations.
Cornell Notes
The paper constructs DESI DR1 LSS catalogs (galaxies and quasars) and the associated random catalogs, weights, and veto masks needed for unbiased two-point clustering measurements. It then validates the resulting configuration- and Fourier-space multipoles against realistic fiber-assignment mocks, finding generally consistent agreement at the 2% level in the inferred overdensity field, with the largest residual issues for ELG and on the lowest Fourier scales.
What is the paper’s central research question?
How to build and validate DESI DR1 galaxy/quasar LSS catalogs (including randoms, veto masks, and weights) so that measured two-point clustering statistics can be modeled accurately despite survey, instrumental, and astrophysical systematics.
What data source and time span are used?
DESI DR1 main survey observations processed with the “iron” spectroscopic reduction, using data through June 14, 2022.
How are the final clustering samples defined for each tracer?
By tracer-specific spectroscopic success criteria (e.g., BGS: ZWARN==0 and ; LRG: ZWARN==0 and ; ELG: ; QSO: not rejected by quasar catalog) plus redshift cuts (BGS , LRG , ELG , QSO primary ).
What are the final sample sizes used for clustering?
BGS 300,043; LRG 2,138,627; ELG 2,432,072; QSO 1,223,391 total, with 856,831 in the primary range.
How does the paper correct for fiber assignment incompleteness?
It defines assignment completeness and uses fiducial completeness weights , with additional handling of tile-group effects via applied to randoms. It also applies a -cut (pairs with ) to avoid small-scale biases.
What is the fiducial imaging-systematics mitigation strategy?
Tracer- and redshift-bin-dependent regressions using Healpix imaging property maps: BGS uses linear regression; LRG uses linear regression; QSO uses random-forest regression; ELG uses SYSNet. The regression-derived inverse models are turned into weights .
How are spectroscopic systematics handled?
By modeling redshift success as a function of TSNR2 (and sometimes fiberflux and redshift) and applying redshift-failure weights .
What two-point estimators are used and in what spaces?
Correlation function multipoles via the Landy–Szalay estimator (with Legendre projections) and power spectrum multipoles via an FKP-based estimator with FFTs.
What validation result do the authors emphasize?
Data and realistic fiber-assignment mocks are generally consistent to within about 2% in the inferred real-space overdensity field; detailed comparisons show the largest residual mismatches for ELG at low and for QSO in Fourier-space quadrupole, improving when restricting to .
Review Questions
Explain why the paper uses a -cut and window-matrix modeling rather than relying solely on completeness weights for fiber assignment.
Which imaging-systematics regression method is used for each tracer (BGS, LRG, ELG, QSO), and how are the resulting weights validated?
Describe the role of randoms in inducing radial integral constraint (RIC) and how the paper corrects for it in Fourier-space.
From Table 9, identify one tracer where configuration-space agreement is strong and one where Fourier-space agreement requires scale cuts; summarize the key behavior.
Key Points
- 1
The paper provides end-to-end DESI DR1 LSS catalog construction: target selection, redshift success cuts, veto masks, random catalogs, and systematic weights.
- 2
Fiber assignment incompleteness is corrected with completeness weights , but unbiased small-scale clustering requires a pair truncation and window-matrix modeling (fiducial approach).
- 3
Imaging systematics are mitigated with tracer-specific regressions (linear for BGS/LRG, SYSNet for ELG, random-forest for QSO) and validated via null tests against uncontaminated mocks.
- 4
Spectroscopic systematics are corrected using redshift-failure weights modeled as functions of TSNR2 (and additional variables for some tracers).
- 5
Two-point statistics are measured as multipoles in both configuration and Fourier space, with explicit corrections for window effects, RIC, and AIC.
- 6
Validation against realistic “altmtl” mocks shows general agreement at the 2% level in inferred overdensity; residual tensions are largest for ELG at low and for QSO Fourier-space multipoles.