1. Introduction
Anthropogenic aerosols are small liquid or solid airborne particles. They are predominantly the secondary result of emissions of aerosol precursor gases emitted via industrial (e.g., sulfur dioxide) processes and are found to have a net negative effective radiative forcing (Bellouin et al. 2020). By cooling the climate, they have offset some of the past greenhouse gas (GHG) induced warming (Foote 1856; IPCC 2021a). It is becoming increasingly important to understand and quantify the individual impacts these two forcings have had on the climate system over the historical period, to improve constraints on observed GHG effects and thus future predictions, given GHG and aerosol emissions will likely follow opposing pathways (Andreae et al. 2005; Lund et al. 2019; Larson and Portmann 2019; Persad et al. 2022).
While GHGs trap (outgoing) longwave (LW) radiation, heating the climate system, aerosol-associated cooling results from aerosols’ interactions with (incoming solar) radiation and clouds, referred to as aerosol–radiation interactions (ARIs) and aerosol–cloud interactions (ACIs; Bellouin et al. 2020). Aerosols directly absorb and scatter incoming solar radiation, causing changes to the energy budget at Earth’s surface. Some aerosols, such as sulfate, can further serve as cloud condensation nuclei (CCN), affecting cloud albedo (first indirect or Twomey effect; Twomey 1977) and cloud lifetime (second indirect effect; Albrecht 1989). The relatively short lifetimes of aerosols, along with a shift in the main emission regions from North America and Europe to Southeast Asia, have led to spatial and temporal heterogeneous aerosol distributions, with their effects being most dominant over the main emission regions and downstream thereof (e.g., Undorf et al. 2018). Thus, regional aerosol-induced cooling has been associated with shifts in large-scale circulation patterns resulting in, e.g., suppressed summer monsoon precipitation in West Africa, the Sahel, and Asia (e.g., Bollasina et al. 2011; Polson et al. 2014).
The most recent IPCC report assesses “the likely range of total human-caused global surface temperature increase” since the mid-twentieth century as 0.8°–1.3°C where “well-mixed GHG contributed a warming of 1.0°–2.0°C,” while “other human drivers (principally aerosols) contributed a cooling of 0°–0.8°C” (IPCC 2021b). This shows that there is more confidence in the influence of combined human emissions on the observed warming than in separating the effects of GHG and aerosols which causes much larger uncertainties. To disentangle the contributions to an observed change of the two different climate drivers from those of internal climate variability, detection and attribution methods are used. These methods commonly assume that observations are comprised of a linear combination of model responses to different forcings, usually referred to as fingerprints plus noise. Using optimized linear regression, as introduced by Hasselmann (1993) and further developed by Hegerl et al. (1997), Allen and Tett (1999), Allen and Stott (2003), and Ribes et al. (2013), the combination best fitting the observations can then be derived by estimating scaling factors for the individual fingerprints. Finally, these estimated magnitudes from observations can be used to constrain future impacts.
Constrained attributable GHG warming (e.g., Bindoff et al. 2013; Gillett et al. 2021) has been estimated by averaging across response patterns from multiple models [multimodel mean (MMM)], which has been found to provide a more robust assessment by smoothing model errors and biases (Hegerl and Zwiers 2011). However, while a range of studies have detected the contribution of GHG, the response to non-GHG anthropogenic forcings (predominantly aerosols) is not as robust, and results using time–space patterns from different climate model simulations yield quite different results compared to each other and the multimodel mean (e.g., Jones et al. 2013; Gillett et al. 2013; Jones et al. 2016; Schurer et al. 2020). The use of the multimodel mean averages across aerosol uncertainty and is likely to underestimate the true uncertainty (Schurer et al. 2018). This is due to a variety of representations of aerosol processes in models (Wilcox et al. 2015) leading to uncertainties in the simulated climate response to historical aerosol forcing in addition to uncertainty in the forcing itself. Studies have shown that these differences are mostly resulting from varying representations of ACIs (Wilcox et al. 2015; Zelinka et al. 2014, 2020) that are essential for reconstructing historical global warming (Wang et al. 2021).
In this study, we focus on observed changes where aerosol influences have been previously detected in the literature, and using large ensembles from phase 6 of the Coupled Model Intercomparison Project (CMIP6; Eyring et al. 2016), we investigate whether combining time series of temperature and precipitation changes can improve the detection of and constraint on aerosol contributions. Recent studies, by Bonfils et al. (2020) looking at a joint change in temperature, precipitation, and continental aridity and Schurer et al. (2018) investigating the transient climate response by merging various temperature indices, show that adding physically motivated spatial and temporal information can improve results of a detection and attribution analysis and improve the constraint on individual forcing contributions. Based on the success of these previous studies, we apply scaling factors derived for joint changes in temperature and precipitation to estimate aerosol contributions to the observed warming since the mid-twentieth century.
The remainder of this paper is organized in the following order. Section 2 describes the framework of detection and attribution methods applied in this study, our selection of fingerprints, and the observed and modeled temperature and precipitation data. Section 3 is split into two parts. First, using the CMIP6 large ensembles, a hydrological sensitivity study and a perfect and imperfect model study are conducted for both individual models and the multimodel mean to test our methodology. In the second part of section 3, we estimate scaling factors for single and joint fingerprints of the response to aerosols and other forcings and apply them to estimate aerosol contributions to observed global warming. The paper concludes with a discussion of our approach and results in section 4.
2. Methods and data
a. Detection and attribution methods
In the joint study, we conduct a single detection and attribution analysis on the joint fingerprint. Joint fingerprints are obtained by concatenating the single-variable fingerprints along the time axis. This means the joint fingerprint [A; B] of the single variables A and B is given by concatenating the time series of A and B, where
b. Choice of fingerprints
To separate aerosol from GHG influences, we choose diagnostics that have been physically motivated in the literature to differentiate between the two. These are based on both the individual climate impacts these two forcings have and the spatially and temporally unique emission and response pathways they started to follow in the late twentieth century. While GHG concentrations are globally homogeneously distributed and have continued to increase globally since the 1850s, aerosol emissions show unique spatial concentration patterns due to their relatively short lifetimes in the atmosphere and shifts in their macro emission regions. Aerosol emissions started to decline in Europe and the United States following the implementation of national air quality legislation in the 1980s. At the same time, emissions in Asia experienced a rapid increase as a result of economic development leading to a shift of emissions from Europe/North America to Southeast Asia (Bellouin et al. 2020; Hoesly et al. 2018; McDuffie et al. 2020). Nevertheless, simulated responses to aerosol forcings vary strongly between models (Wilcox et al. 2015), which leads to differences in fingerprints for AER from different models and thus model sensitivity of detection and attribution results. Finally, as aerosols impact both temperature and precipitation through direct and indirect effects, with aerosol responses expected to be substantial (Allen and Ingram 2002), we are considering both temperature- and precipitation-based variables to investigate aerosol impacts.
The first fingerprint we investigate is the GMT due to its large signal-to-noise ratio (see Fig. 1a) which has been shown to work well to detect anthropogenic climate change signals but shows large uncertainties when trying to use it alone to distinguish between aerosol and GHG impacts (Wilcox et al. 2013; Schurer et al. 2018; Gillett et al. 2021). Comparing the climate response to AER and noAER forcing in Fig. 1a shows that noAER (largely GHG) warming has likely been offset to some extent by AER-induced cooling. The strongest aerosol signal was found until the 1980s when global aerosol emissions were still increasing.
Aerosol emissions being predominantly located in the Northern Hemisphere (NH) have induced a change in the interhemispheric temperature asymmetry (ITA; Friedman et al. 2013; Wilcox et al. 2013; Schurer et al. 2018) defined as the difference in temperature between NH and Southern Hemisphere (SH). Model studies suggest that until the 1980s, the ITA response to aerosol and GHG forcing offset each other; then GHG effects start to dominate due to declining global aerosol emissions (Friedman et al. 2013). This shift is represented by a sharp increase of the ITA due to a disproportional GHG warming of the Arctic and NH landmasses which had previously been compensated by aerosol-induced NH cooling (Wilcox et al. 2013; Friedman et al. 2013, 2020; Shindell et al. 2015). The change in the trend of the ITA in the CMIP6 large ensembles and observations (Fig. 1b) reflects aerosols’ emission history clearly showing a reduction in the ITA in the AER and ALL signals and the observations until the mid-1980s followed by a sharp increase. The spread between CMIP6 models is larger for the ITA than GMT, highlighting model uncertainties in simulating (regional) aerosol impacts (Wilcox et al. 2015).
Changes in global mean temperatures are considered to drive thermodynamic changes in the hydrological cycle described through the Clausius–Clapeyron relationship (Allen and Ingram 2002; Hegerl et al. 2015) and dynamic changes by shifting large-scale atmospheric circulation, such as the southward shift of the intertropical convergence zone (ITCZ; Schneider et al. 2014). Opposing the increase in precipitation associated with GHG warming, aerosol cooling has, therefore, been found to result in a reduction of evaporation and precipitation (Ming and Ramaswamy 2009; Bony et al. 2013; Polson et al. 2013b), especially over land where sulfate particles both cool the surface and serve as CCN (Ramanathan et al. 2001; Zhang et al. 2021). Given that instrumental records on land go back further than the satellite era (1980s), we investigate global mean land precipitation (GMLP) as a diagnostic. ALL, noAER, and AER signals for the GMLP are shown in Fig. 1c and highlight aerosol-induced drying until around 1990 and GHG-related wetting. Again, the ALL fingerprint and observation suggest that both forcings have been working against each other.
In the tropics, the intensification of the water cycle due to global warming has led to a “wet-get-wetter, dry-get-drier” pattern when tracking the motion of wet and dry regions (e.g., Held and Soden 2006; Polson et al. 2013a). Schurer et al. (2020) observed that the opposite is the case for aerosols. By cooling the climate system, models suggest a drying of the wet regions and a wetting of the dry regions associated with increasing global aerosol emissions. Thus, following a method developed by Polson et al. (2013a) and Polson and Hegerl (2017) for each monthly precipitation field, we rank tropical (30°S–30°N) grid boxes according to their absolute monthly total precipitation in ascending order. The upper 30% are designated as wet regions, and the bottom 30% are designated as dry regions. The time series of average monthly precipitation from area-weighted means of the respective regions is shown in Figs. 1d and 1e and shows clear changes in wet regions, while the impact from AER and noAER is less clear for dry regions. As changes in both regions are linked through the transport of water, we will treat tropical wet and dry regions as a single fingerprint “wet dry” (WD), where the individual time series are concatenated along the time axis to form a single fingerprint (Schurer et al. 2020).
Other potential climate variables were investigated before choosing those mentioned above, such as the diurnal temperature range (DTR; Stjern et al. 2020) defined as the difference between daily maximum and daily minimum temperature. Due to GHG and aerosols’ unique impact on radiation, the DTR has been suggested as a diagnostic to separate the response to the two forcings (e.g., Wild et al. 2007; Stjern et al. 2020; Undorf et al. 2018). However, as we found large differences between models and a low signal-to-noise ratio, we decided against including the DTR as a potential fingerprint in this study. Referring to shifts in aerosol macroemission regions from the Western Hemisphere to the Eastern Hemisphere, we also conducted a meridional analysis of aerosol impacts on ITA and tropical wet and dry regions, but our investigation did not return a robust spatial signal of the shift to more southeastern emission regions in either temperature or precipitation reemphasizing the impact of model uncertainty.
c. Observations and climate model data
Full coverage of the tropics is required to track wet and dry regions; thus, following Schurer et al. (2020) and the IPCC (2021a), the satellite–gauge merged dataset of monthly precipitation data from the Global Precipitation Climatology Project version 3.2 (GPCPv3.2; Huffman et al. 2023) is used. Although observations are available from January 1979, the analysis is limited to the period January 1988–December 2020, where measurements from the Special Sensor Microwave Imager are available (following Polson and Hegerl 2017). A sensitivity test and a comparison of different datasets for tropical wet and dry regions can be found in Schurer et al. (2020). To calculate GMLP for the period 1955–2020, the Climate Research Unit’s gridded Time Series version 4.05 (CRUTSv4.05; Harris et al. 2020) dataset, for which monthly observations are available starting in January 1901, is used. Other observational precipitation datasets are available, which differ in their distribution of station input, homogenization, area averaging, and quality control sampling. However, it has been argued that the homogeneous coverage of CRUTS provides a more reliable time series (Lorenz and Kunstmann 2012). Finally, observed changes in the ITA and GMT are calculated from monthly temperature anomalies in the Hadley Centre/Climatic Research Unit, version 5 (HadCRUT5; Morice et al. 2021), noninfilled dataset, where data are available from January 1850 onward.
We limit our detailed analysis to models with at least six ensemble members, i.e., large ensembles. Thus, large ensembles of three CMIP6 models are analyzed (CanESM5, CNRM-CM6-1, and IPSL-CM6A-LR) for which CMIP6 historical, Scenario Model Intercomparison Project (ScenarioMIP) Shared Socioeconomic Pathway (SSP) 2-4.5, and Detection Attribution Model Intercomparison Project (DAMIP) hist-aer simulations of monthly near-surface air temperature and precipitation are available. As SSP2-4.5 forcings are used in DAMIP simulations for the period 2015–20, we merge CMIP6 historical simulations with SSP2-4.5 simulations for 2015–20 to be able to extend the analysis to 2020. Note that because HadCRUT5 contains temperature anomalies relative to 1961–90, for consistency, we take monthly anomalies of each grid box in reference to 1961–90 in the temperature simulations. This is done before GMT and ITA fingerprints are derived. While testing of the methodology in section 3a is purely based on these large ensembles, in order to compare the large ensembles to the remaining CMIP6 model ensemble, we additionally download model ensembles for which at least three historical, historical SSP2-4.5, and DAMIP simulations are available (see Table 1). The multimodel mean of the total six CMIP6 models will be analyzed as a reference fingerprint, referred to as “CMIP6 MMM,” in the observational analysis in section 3b.
Models used in this analysis. CMIP6 models are selected from the CMIP6 cohort if CMIP6 historical, SSP2-4.5, and DAMIP single forcing simulation for temperature and precipitation are available. Aerosol schemes are separated into specified (externally simulated fields of aerosol optical depth) and prognostic (propagating physical aerosol properties) schemes. The second column shows the number of ensemble members included in each model ensemble. The blank row highlights the separation between the CMIP6 large ensembles used in section 3a (top half) and the remaining CMIP6 models with at least three ensemble members (bottom half) which are analyzed as an additional reference fingerprint in section 3b. Values for the ECS are taken from Gettelman et al. (2019) and Meehl et al. (2020), and values for ERFAERnet of the net aerosol effect (ERFAERnet = ERFACI + ERFARI) are taken from Zelinka et al. (2014) and Smith et al. (2020). IPCC estimates a very likely range of 2°–5°C for ECS (Forster et al. 2021). (*) Aerosol ERF was calculated using a prognostic aerosol scheme. While all models include the first indirect effect, the Twomey effect (Twomey 1977) models also simulating the second indirect effect are denoted with two asterisks (**).
For consistency, all observational datasets and the model data are regridded to a 2.5° × 2.5° grid. For precipitation (i.e., CRUTS, GPCPv3.2, and precipitation model simulations) a conservative regridding method is chosen (preservation of the integral of the precipitation field; Jones 1999), while temperature fields (HadCRUT5 and temperature model simulations) are bilinearly regridded, i.e., linear interpolation in two directions using the four nearest grid boxes. Because CRUTS and HadCRUT5 only contain values for grid points where observations are available, fewer measurements exist for earlier decades. To compensate for this, we derive observational masks from the regridded observational datasets to only use grid cells for which at least one observation in each season in each year (i.e., 4 per year) is provided over the period 1955–2020. The CRUTS and HadCRUT5 seasonal masks are then applied to both the respective observations and model data to calculate the fingerprints (see Table 2). Then 3-yr means of the fingerprints are taken, and both the observations and the model simulations (including the control samples) are standardized. We standardize the time series by dividing the fingerprints by the average standard deviation of the 500 control simulation samples from the full CMIP6 model ensemble for each respective climate variable. These are the same control simulations used during the detection and attribution analysis (see methods in section 2). The average standard deviation is obtained by taking the mean of the 500 standard deviations of each individual control simulation. Standardizing the time series accounts for the relative difference in magnitude of the change in the different temperature- and precipitation-based fingerprints, and this allows us to avoid the larger magnitude of, for example, GMT, dominating the analysis (Schurer et al. 2020). In the final step, the fingerprints are centered, i.e., anomalies are taken over the whole analysis period.
Fingerprints investigated in this analysis. Variable names with the respective abbreviations, the dataset used to derive the time series, the analysis periods, and the observational mask applied to calculate the time series are listed.
d. Aerosol forcing in CMIP6
Aerosol schemes have been updated in this newest generation of climate models compared to previous versions. This concerns especially aerosol microphysics and ACIs with all models in this study containing direct effects and at least the first indirect effect, i.e., cloud albedo effect or Twomey effect (Twomey 1977). Studies have found that CMIP5 models that simulate some form of the Twomey effect reproduce representations of interdecadal variability in historical GMT, land-only precipitation, and zonal temperature trends better than models that only include the direct effect (Wilcox et al. 2013; Ekman 2014). However, radiative feedbacks from clouds, especially ACIs, have also been associated with causing a large spread in predictions of future global warming (Boucher et al. 2013; Zelinka et al. 2020; Smith et al. 2020). Stronger cloud feedbacks have further been linked to causing an increase in spread and magnitude of equilibrium climate sensitivity (ECS), a standard metric to measure model sensitivity to carbon dioxide (CO2) increase, among CMIP6 models although some disagreement remains on the correlation between effective radiative forcing from aerosols (ERFAERnet) and ECS in CMIP6 models (Meehl et al. 2020; Smith et al. 2020). Similarly, no significant association between stronger aerosol forcing and higher ECS is found for CMIP5 models when considering the range of models (Forster et al. 2013; Chylek et al. 2016). However, this is different when only considering CMIP5 models that include aerosol indirect effects (Chylek et al. 2016). While models of both high and low ECS are able to simulate the observed evolution of GMT, inconsistencies arise when investigating the representation of observed interhemispheric temperature records, where high ECS models produce a strong ACI cooling in the NH (Wang et al. 2021). Nevertheless, the models considered in this study have been found to be consistent with recent observational estimates of ERFAERnet (from −2.0 to −0.4 W m−2 as a 90% confidence interval; Bellouin et al. 2020).
Further differences between models arise from varying aerosol schemes with some models having specified schemes, where 3D aerosol fields are simulated externally before being used as aerosol forcing in the CMIP6 simulations, while others include prognostic aerosol schemes propagating physical aerosol properties within the model (see Table 1). These differences between aerosol schemes have been shown to have the highest impacts in regions remote from aerosol emission sources due to differing dynamical responses partially driven by deviating aerosol optical depths (AODs), especially over oceans. This is likely due to relative humidity-related aerosol swelling in prognostic models leading to increased aerosol optical depth and additional cooling (Randles et al. 2013).
3. Results and discussion
This results section is separated into two subsections using different methodologies and data. A schematic describing the content of section 3 is displayed in Fig. 2. In the first subsection (section 3a), we test our approach of joining temperature and precipitation fingerprints using the CMIP6 large-ensemble model simulations. The method validation is conducted in four steps, analyzing the models’ temperature and precipitation sensitivities to external forcings and the ability of our choice of fingerprints to separate AER and noAER contributions in model data (“imperfect model study”). In the remainder of this section (section 3b), we include observations and the CMIP6 MMM reference fingerprint to detect and attribute AER and noAER influences on observed climate changes. Finally, to evaluate the implications of combining temperature and precipitation fingerprints, we plot the mean 2010–19 minus mean 1850–1900 temperature anomalies of scaled model GMT fingerprints for AER and noAER. The fingerprints are scaled using scaling factors from the joint study.
a. Testing of the methodology using CMIP6 large ensembles
1) Should we combine temperature and precipitation fingerprints?
If the sensitivity to an external forcing in the model is incorrect and the model fingerprints need to be scaled to match the observations, then we are assuming that both the temperature and precipitation fingerprints need to be scaled by the same amount. Several reasons exist why this is not always true, such as the model having the incorrect sensitivity to external forcings, a wrong forcing (which can also be due to the observations serving as forcing input), or simulating the change of a pattern wrong. Global changes in precipitation are linked to changes in temperature referred to as hydrological sensitivity (HS) which is defined as the relative change in global mean precipitation (GMP) as a function of degree change in global mean temperature (% K−1). The HS is constrained by the energetics of the atmosphere to 2% K−1 (e.g., Allen and Ingram 2002; Hegerl et al. 2015).
Using similar observational datasets to those used in this study, Allan et al. (2020) derive an observational HS of 2% ± 0.5% K−1 for the period 1979–2019, while model-based studies (e.g., Norris et al. 2022; Fläschner et al. 2016) have found a wide range of HS values with large scatter arising from the choice of the studied period and the chosen dataset. Norris et al. (2022) calculate a model spread in HS of 1.1%–2.2% K−1 using 150 years of CMIP6 1pctCO2 simulations (increase of CO2 by 1% yr−1 from preindustrial levels), while comparing CMIP6 abrupt carbon dioxide quadrupling (abrupt-4xCO2) experiments to piControl simulations, Pendergrass (2020) find a range of HS of 2.1%–3.1% K−1 in the CMIP6 ensemble. Using historical ALL simulations, we estimate a range of best estimate HS values for the models used here of 1.5%–1.71% K−1 for the period 1988–2020 and slightly lower values ranging from 1.21% to 1.38% K−1 when extending the period to 1955–2020 (see Fig. 3). Lower HS values could be due to the stronger aerosol forcing in the longer period, which overcompensates the GHG-related precipitation increase potentially due to ACIs, thus decreasing the HS. Previous studies have noted that aerosols might impact precipitation more than GHG (e.g., Allen and Ingram 2002; Hegerl et al. 2015). Thus, our model results (1988–2020) lie in the lower bounds of the observed HS (1979–2019) and align with model-based studies although we find lower values also likely due to historical aerosol forcing. As our models lie in the uncertainty range of the observed HS, we assume that the fingerprints can be combined.
2) A perfect model study of our choice of fingerprints
To evaluate our choice of fingerprints in their ability to detect aerosol influences, a perfect model study is conducted, i.e., to analyze the signal-to-noise ratio of our choice of single-variable fingerprints (Ribes and Terray 2013; Schurer et al. 2018; Gillett et al. 2021). Here, one of the ALL ensemble members is used as a pseudo-observation, and all other ALL and AER members from the same model are used to derive fingerprints to determine scaling factors βALL for a 1-signal study and βnoAER and βAER for a 2-signal study. This process is repeated for all historical model simulations for all models. By definition, this analysis should give a 5%–95% confidence interval where on average 1 is included in the scaling factor range in 90% of cases allowing us to evaluate variables in their ability to reconstruct observed changes as well as separate noAER and AER contributions. To rate the individual variables, we calculate the percentage where the scaling factors deviate from 1 (fail to include 1) referred to as the deviation rate. Note that cases where uncertainties cannot be constrained, i.e., infinite confidence intervals, are also assumed to include 1.
Results from the perfect model study are shown in Fig. 4. Across variables, we find the lowest uncertainties for βALL, while separating the noAER (βnoAER) and AER (βAER) contribution leads to less constrained scaling factors with the largest uncertainties observed for βAER. We find the smallest deviation rates for GMT and GMLP depending on the model. However, our analysis shows that across models, GMT is the best single diagnostic to attempt to separate AER and noAER contributions, as it leads to the tightest constraint. There is more uncertainty and variability between models in separating AER and noAER contributions for WD and ITA likely due to varying aerosol schemes between models. Among the three large ensembles, CanESM5 is the only model with a prognostic aerosol scheme, simulating a strong cooling induced by aerosol impacts on clouds in the Northern Hemisphere which results in a clear AER trend in the ITA, GMT, and GMLP, while less of an aerosol signal is visible for CNRM-CM6-1 and IPSL-CM6A-LR (see Fig. S1 in the online supplemental material). The weaker aerosol signal might, therefore, be due to the different aerosol schemes although the reduced ensemble size of CNRM-CM6-1 and IPSL-CM6A-LR (10 and 6, respectively) compared to CanESM5 (30) could also be a factor. While CNRM-CM6-1 has a lower deviation rate when separating noAER and AER impacts for WD than ITA in the perfect model study, the opposite is the case for CanESM5. Note that the unsuccessful attribution of the WD response to noAER and AER forcing in IPSL-CM6A-LR is likely due to a low signal-to-noise ratio (cf. Fig. S1). Although, ideally, we would expect to find deviation rates of around 10% (i.e., 90% coverage of the confidence intervals), this is not the case for all models, especially CNRM-CM6-1 when separating AER and noAER contributions (20%–30% in Fig. 4). Studies mentioned above (e.g., Li et al. 2021) have noted this problem before. This could be a combination of underestimated uncertainty ranges using TLS when the signal-to-noise ratio is weak, such as in the precipitation fingerprints (DelSole et al. 2019; Li et al. 2021; Ma et al. 2023), and because we sample control simulations from the range of CMIP6 models. Therefore, scaling factors of models with larger internal variability than the sampled control variability (CNRM-CM6-1 and IPSL-CM6A-LR; Parsons et al. 2020) might return overconfident results (Schurer et al. 2018) and reversed.
3) An imperfect model study using single variables
An imperfect model study further highlights the varying model sensitivities to external forcings (Fig. 5; note that colors depict the model from which pseudo-observations are taken) as it differs from a perfect model study by using one of the ALL simulations from one model as pseudo-observations while fingerprints are derived as the multimodel mean of the remaining models. Thus, we compare model responses to each other, and scaling factors in the imperfect model framework, therefore, account for model differences, including in the aerosol response, and highlight the varying sensitivities to external forcings between the different models (Zelinka et al. 2020). For this reason, we also do not expect to find deviation rates of 10%, i.e., the scaling factors should be different to 1, and uncertainty in the aerosol forcing itself and model uncertainty contribute to the scaling factor uncertainties (e.g., Wilcox et al. 2015). As for the perfect model study, we find the lowest uncertainties for GMT across all models, where among the different scaling factors, βALL is constrained the most and βAER is constrained the least. As expected, the range of scaling factors widens to account for model differences (see Fig. 5). Overall, consistency in scaling factors for GMT, ITA, and GMLP across models indicates that the models show a comparable response to external forcings in both temperature and precipitation, further supporting our approach of combining the two fields in a joint study. As for the perfect model study, we observe an undercoverage of the estimated confidence intervals (e.g., Li et al. 2021).
4) An imperfect model study of the joint fingerprints
An imperfect model study of combined fingerprints evaluates the reliability of our approach and its sensitivity to model differences in fingerprints. By combining the time series in a joint study, we add physically motivated temporal and spatial information, which supports the separation of aerosol cooling from GHG (noAER) warming. We investigate seven different combinations of diagnostics: two fingerprints combining precipitation ([WD; GMLP]) and temperature ([ITA; GMT]) variables and fingerprints summarizing the global impact ([GMT; GMLP]), dynamical shifts ([ITA; GMLP]), and thermodynamic changes ([WD; GMT]). We then add information to the dynamical fingerprint ([ITA; GMT; GMLP] and [WD; ITA; GMT]), before joining all variables ([WD; ITA; GMT; GMLP]). The joint study is conducted by concatenating the respective fingerprints and conducting a single analysis on the joint fingerprint (see section 2).
In Fig. 6, estimates of the pseudo-observed AER and noAER contributions to changes in GMT (2010–19 minus 1850–1900) from model simulations are shown (y axis) to diagnose performance in the imperfect model study. Scaled estimates are constructed by multiplying scaling factors from the imperfect model study (see Fig. S3) with the respective AER and noAER GMT multimodel mean fingerprints (from two models) and then by subtracting the 1850–1900 mean from the mean for 2010–19 (see methods in section 2). These attributed contributions from the imperfect model study are then compared to the (true) simulated AER and noAER GMT contributions derived from the GMT model mean of the model serving as pseudo-observations (x axis). Finally, to evaluate the skill of the various diagnostics, the percentage agreement between attributed and true contributions is calculated, referred to as the success rate. As we find that GMT is the best diagnostic to separate AER and noAER contributions (see Fig. 4 and Fig. S2d) among the single variables, we are verifying our approach by comparing it to results obtained for GMT-only shown in Fig. 6a (same as in Fig. S2). Figures 6b–h display attributed contributions for an imperfect model study of the combined diagnostics. Here, the color of the scatter points depicts the model that serves as pseudo-observation from which the (true) simulated contribution is estimated.
Our assumption of similar model sensitivities to external forcings for both temperature and precipitation seems to hold as we still observe stronger simulated GMT contributions of AER and noAER when CanESM5 is used as pseudo-observations across combined diagnostics (orange circles). We find that our approach shows some skill in improving the attribution of AER contributions. Across models and combined diagnostics, the fraction of unconstrained scaling factors decreases compared to individual variables (only a single unconstrained case for [WD; GMLP] and [WD; ITA; GMT; GMLP]). Further, for the majority of combined diagnostics, the percentage of cases including the true simulated value of the AER contribution is similar to GMT on its own (success rate AER ≥ 86%). Although attributed contributions are less constrained in some cases (e.g., [WD; GMLP]), similar uncertainties to GMT-only are observed for AER and noAER contributions when including both global means ([GMT; GMLP] and [ITA; GMT; GMLP]). In general, whether the addition of the wet–dry contrast or the interhemispheric temperature asymmetry to the combined fingerprints increases or decreases the uncertainty of estimated AER and noAER contributions is model dependent.
However, our study also suggests that combining fields introduces a tendency to overestimate noAER contributions, i.e., estimates lie above the 1:1 line. As pseudo-observations with higher variability than the fingerprints (and the piControl simulations to calculate the covariance matrix) will have overconfident results, i.e., small confidence intervals, the combination of overestimation and small uncertainty ranges leads to success rates of less than 90% (especially for noAER). Among models in this study, CNRM-CM6-1 and IPSL-CM6A-LR have very high internal variability compared to most of the remaining CMIP6 models and CanESM5 (Parsons et al. 2020). Finally, as mentioned before, model uncertainty and undercoverage of the confidence intervals arising from the TLS analysis can also play a role in observing success rates of less than 90% in an imperfect model study (Schurer et al. 2018; DelSole et al. 2019), particularly where noisier diagnostics are used (e.g., [WD; GMLP]). However, we note that the signal-to-noise ratio tends to increase when we combine variables.
b. Detection and attribution of observed climate changes
Regressing the ALL fingerprint onto the observations (1-signal analysis; βALL) confirms that external forcings play a role in driving observed changes across variables (βALL > 0). Scaling factors for the three individual CMIP6 large-ensemble models are compared in Fig. 7, along with their multimodel mean (LE MMM; black). For additional comparison, a larger CMIP6 multimodel mean (CMIP6 MMM; gray) is also derived containing the six CMIP6 models listed in Table 1. The respective observational datasets for individual single variables are listed in Table 2. Note that while in previous plots colors depict the model serving as a source for the pseudo-observations, they now indicate the model from which fingerprints are derived.
Comparing βALL for temperature- and precipitation-based variables, it becomes clear that across models, precipitation-based variables show a larger scaling factor, i.e., smaller change compared to the observations, than those for temperature. For CNRM-CM6-1 and IPSL-CM6A-LR, ITA and GMT scale around 1 (Figs. 7c,d) and are thus consistent with the observations, while scaling factors for WD are well constrained and lie around 2 (Fig. 7a). For GMLP, scaling factors are higher (between 2 and 3) and less constrained. CanESM5, which exhibits a stronger response to external forcings (see Fig. S1), overestimates the ITA and GMT responses but shows scaling factors around 1 for WD and GMLP. The large-ensemble multimodel mean matches the mean responses of the three individual models and is consistent with the CMIP6 multimodel mean. Uncertainties are higher for CNRM-CM6-1 and IPSL-CM6A-LR, as they consist of smaller ensembles.
AER and noAER contributions are detectable in a 2-signal analysis for the interhemispheric temperature asymmetry and global mean land precipitation (βnoAER and βAER > 0; Fig. 7), but a residual consistency test shows that for CanESM5 and CNRM-CM6-1, GMLP residuals are larger than the 95th percentile of internal variability estimates for GMLP indicating large unexplained variability. They are surprisingly small for ITA (<5th percentile). Thus, the scaled noAER and AER fingerprints for GMLP explain less observational variability than expected indicating a possible underestimation of precipitation variability in the sampled control simulations. At the same time, smaller residuals for ITA suggest a possible overestimation in the hemispheric temperature variability in the control simulations. GMT responses to AER and noAER forcings are attributable to IPSL-CM6A-LR only.
1) Why do we observe different scaling factors for temperature and precipitation when using observations?
By comparing scaling factors for temperature and precipitation in the imperfect model study and from the hydrological sensitivity analysis, we would assume that the analyzed models exhibit the same sensitivity in temperature and precipitation to external forcings. However, scaling factors from the 1-signal analysis (βALL) in Fig. 7 suggest that temperature and precipitation scale differently. To analyze if this is true or if the differences are consistent with uncertainty, we conduct an observational detection and attribution study for the same 2-model mean fingerprints as in the imperfect model study. The resulting scaling factors for WD (y axis) are plotted against temperature-based (x axis) variables for the observational study (black whiskers) and the imperfect model results (colored whiskers) in Fig. 8. Since this difference in scaling factors is especially true for WD, only results for WD are shown here. Scaling factors for GMLP are more consistent with those from GMT and ITA and are displayed in Fig. S6. Comparing scaling factors for temperature and WD indicates that the models do not exhibit the same response in temperature and precipitation to historical forcings, as results do not fall on the diagonal (dotted line). In general, the best estimates of the precipitation-based variables (cf. GMLP in Fig. S6) are closer to 1, tighter constrained, and show more consistency with ITA and GMT when CanESM5 is included in the fingerprint (purple and green scatter points) confirming that a stronger precipitation response to external forcings is needed to reconstruct observed changes. This inconsistency between observational and imperfect model results raises the question of whether these differences are within the range of internal model variability. Previous studies have found discrepancies between simulated and observed precipitation variability in climate models (e.g., Zhang et al. 2007; Hegerl et al. 2015), particularly in the tropics. Thus, we continue this analysis with doubled estimates of model precipitation variance (as it has been done in, e.g., Polson et al. 2013a; Schurer et al. 2020) to address the failure of variance, since we observe an underestimation in precipitation variability when separating noAER and AER influences on GMLP in the previous section (see Fig. 7b), residuals that are larger than expected and our models are consistent with observational hydrological sensitivity. This also leads to a decrease in residuals for WD and GMLP (see Fig. S4).
Overall, results from the imperfect model study suggest that it is helpful to combine precipitation and temperature fingerprints for a clearer estimate of the climate response to external forcings.
2) Can we improve the detection and attribution of aerosol influences in observations?
Based on imperfect model results, the combination of global means in [GMT; GMLP] could be considered the most reliable to estimate aerosol contributions to observed changes as the joint diagnostic estimates the “true” AER contribution with a success rate of close to 90% (similar to GMT; see Fig. 6) and shows the tightest constraint. To confirm this, forcing contributions to the observed warming in 2010–19 relative to 1850–1900 using scaling factors from an observational joint study (see Fig. S5) are derived, and for the large-ensemble multimodel mean (LE MMM) plotted in Fig. 9 (for individual models and the CMIP6 multimodel mean, see Fig. S7). As for the joint study (Fig. 6), contributions are estimated by scaling GMT (multi)model mean fingerprints (covering the period 1850–2019) for AER and noAER with the respective scaling factors (Fig. S5) before subtracting the 1850–1900 mean from the 2010–19 mean. Although the scaling factors are calculated on a shorter period (1988–2020/1955–2020), estimating GMT this way allows us to directly compare and evaluate our approach with recent results from Gillett et al. (2021).
In the case of the large-ensemble multimodel mean, findings from the imperfect model study are supported. The most constrained estimates of AER and noAER contributions of the seven diagnostics analyzed are obtained for [GMT; GMLP] with an aerosol cooling of 0.46 K ([−0.86, −0.05] K) offsetting 1.63 K ([1.26, 2.00] K) of noAER/GHG warming (Table 3). [GMT; GMLP] is slightly improving the detection relative to GMT alone by reducing the uncertainty range of the noAER (mostly GHG) contribution and attributing an AER cooling (<0). We further do not constrain the AER impact on cooling based on GMT alone. The addition of precipitation in the fingerprint, therefore, allows us to attribute a response to aerosol forcing of the sign expected from models in observations. This is consistent with results from the literature that show that AER-induced surface cooling triggers a stronger response in precipitation compared to GHG (Allen and Ingram 2002).
The best estimate contributions and respective 5%–95% uncertainty ranges in squared brackets of AER and noAER to the observed warming (2010–19 relative to 1850–1900) using scaling factors for the LE MMM (MMM of the three CMIP6 large ensembles).
Similar results to those for the large-ensemble multimodel mean are obtained for the individual models and the CMIP6 multimodel mean. As previously noted for the imperfect model study and also here for observations, diagnostics including global means behave best across models, while it is model specific which nonglobal/dynamical variables cause an increase in uncertainty when added to the combined fingerprint. For example, the addition of the interhemispheric temperature asymmetry to [GMT; GMLP] decreases the uncertainty of the AER contribution for CNRM-CM6-1 and IPSL-CM6A-LR, while it increases the confidence interval for CanESM5 due to its stronger AER response (Fig. S7). As for the large-ensemble model mean, AER and noAER contributions can only be clearly attributed for the single-model analyses when using combined fingerprints. We again note that aerosol pattern uncertainty and the undercoverage of uncertainties when using TLS possibly render our results slightly overconfident (Schurer et al. 2018; DelSole et al. 2019).
Apart from [WD; GMLP] and [WD; ITA; GMT] fingerprints, large-ensemble estimates for noAER and AER agree with recent findings from a 3-signal study in Gillett et al. (2021) which attributes contributions of GHG+, NAT, and AER as [1.2, 1.9], [−0.01, 0.06], and [−0.7, −0.1] °C, respectively. Although we are using a similar methodology, with the main difference being that we do not consider observational uncertainty from the ensemble spread of HadCRUT4 or uncertainty arising from the ratio of global surface air temperature and global mean surface temperature, we find slightly larger uncertainties than Gillett et al. (2021). This is likely due to the fact that attributions in Gillett et al. (2021) are estimated based on the multimodel mean of 13 models and temperature data for the period 1850–2019. For individual models, Gillett et al. (2021) cannot clearly attribute the AER signal when looking at GMT alone, which is consistent with our results. Thus, while the skill of combining temperature and precipitation fingerprints is not so clear for the multimodel mean, it allows for the attribution of aerosol impacts in individual models.
In summary, this study shows that diagnostics combining both precipitation and temperature outperform or perform as well as diagnostics where individual fields of temperature and precipitation are combined ([WD; GMLP] and [ITA; GMT]). Further, our results suggest that a joint study improves the separation of anthropogenic forcing contributions to observed changes relative to GMT on its own. The addition of a precipitation-based fingerprint increases the detectability of the AER forcing decreasing the probability of no contribution (scaling factor equal to zero; except for [WD; GMLP] and [WD; ITA; GMT]). Thus, for most models and diagnostics, we attribute an AER-induced cooling with consistent best estimates of around 0.5–0.8 K and a GHG warming of around 1.7 K.
4. Conclusions
Using large ensembles from three CMIP6 models and observational datasets for temperature (HadCRUT5) and precipitation (CRUTS and GPCP), we comprehensively investigated whether combining physically motivated information from temperature and precipitation into a joint study can reduce the confidence intervals of aerosol contributions to observed changes in GMT. Both perfect and imperfect model studies were conducted to evaluate model uncertainties and rate our choice of diagnostics. We then compared model sensitivities in temperature and precipitation to evaluate our approach before reconstructing aerosol influences and contributions from the remaining forcings (noAER; mostly GHG) to the observed warming. Finally, we compared our results with estimates from a detection and attribution study using GMT-only and values from recent literature, i.e., Gillett et al. (2021), which is based on a similar method, but using multimodel mean fingerprints and temperature-only diagnostics. The main findings from the present study are as follows:
-
A hydrological sensitivity analysis shows that the three large ensembles simulate the lower bound of the observational energetic relationship between GMP and GMT of 2 ± 0.5% K−1 (Allan et al. 2020) and the assumed consistency in model sensitivities in temperature and precipitation to external forcings is further supported by results from an imperfect model study. However, large residuals from a 2-signal analysis for GMLP and inconsistencies between findings from the imperfect model and observational study raise the question of whether our models underestimate (tropical) precipitation variability, as has been found in previous studies (e.g., Zhang et al. 2007; Schurer et al. 2020). To address this, we thus double precipitation variance for the observational detection and attribution analysis.
-
Apart from the purely precipitation-based diagnostic [WD; GMLP], estimates of noAER and AER contributions agree with findings from the literature (Gillett et al. 2021). Our best estimate using the multimodel mean of the three large ensembles is an AER cooling of 0.46 K ([−0.86, −0.05] K 90% confidence interval) and a warming associated with noAER of 1.63 K ([1.26, 2.00] K) using the concatenated time series of GMT and GMLP. We detect negative AER contributions for a range of diagnostics even if GMT is not included and can slightly reduce the uncertainty range of the AER and noAER contributions compared to our study of using GMT alone (AER: −0.38 [−0.81, 0.06] K and noAER: 1.59 [1.21, 1.97] K). Highlighting model sensitivity, we find quite consistent estimates of AER cooling from observations across the three large ensembles: for CNRM-CM6-1 −0.57 [−1.16, −0.01] K, IPSL-CM6A-LR −0.5 [−0.94, −0.09] K, and CanESM5 AER −0.42 [−0.8, −0.04] K. This cooling is counteracted by similarly consistent estimates of noAER across the same respective models of 2.2 [1.67, 2.74] K, 1.95 [1.46, 2.46] K, and 1.18 [0.9, 1.45] K. These estimates are more robust across models than using GMT-only.
-
We find model dependency in the ability of nonglobal mean fingerprints to improve the detectability of the AER signal, i.e., the separation from 0, which is likely due to varying aerosol schemes and aerosol sensitivities across models. This highlights the need to better constrain the spatial patterns of the aerosol response in climate models.
-
Although statistically we would expect confidence intervals to deviate from 1 in only 10% of cases for a perfect model study, we find deviation rates range from 0% to 40%. There are a number of factors contributing to deviations from 10%, such as a low signal-to-noise ratio (especially for the precipitation-based fingerprints alone), and because we sample control simulations from the range of CMIP6 models, which do not represent the true internal variability of the models investigated in this study. This causes both unexpectedly low and high failure rates depending on a model’s true internal variability. In the imperfect model study, we do not expect to find deviation rates of 10% as we are calculating scaling factors for model simulations with different sensitivities to forcings; thus, the scaling factors should be different from 1 (deviation rates range from 0% to 100%). Additionally, model uncertainties due to differences in aerosol forcing among the different models contribute to the uncertainties (Wilcox et al. 2015; Schurer et al. 2018). A final factor to consider is the potential undercoverage of the confidence intervals derived using TLS, which is an intrinsic limitation of the method that becomes apparent for low signal-to-no-noise ratios (e.g., single precipitation or ITA fingerprints). This has been noted by DelSole et al. (2019), Li et al. (2021), and Ma et al. (2023), and our study, further, contributes to highlighting these methodological constraints and the need for future improvements.
Acknowledgments.
This work was supported by the European Union’s Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie Grant Agreement 860100 (iMIRACLI). G. C. H., A. P. B., and A. P. S. were also funded by the Natural Environment Research Council (NERC) project GloSAT (NE/S015698/1). A. P. S. further received funding from a chancellor’s fellowship at the University of Edinburgh. We acknowledge the World Climate Research Programme’s Working Group on Coupled Modelling, which is responsible for CMIP, and we thank the climate modeling groups for producing and making available their model output (listed in Table 1). We thank the iMIRACLI cohort for their guidance and general support throughout the project.
Data availability statement.
Observational data from the Global Precipitation Climatology Project, version 3.2 (GPCPv3.2; Huffman et al. 2023), the Climate Research Unit’s gridded Time Series, version 4.05 (CRUTSv4.05; Harris et al. 2020), and the Hadley Centre/Climatic Research Unit, version 5 (HadCRUT5; Morice et al. 2021), are freely available online (see https://disc.gsfc.nasa.gov/datasets/GPCPMON_3.2/summary, https://catalogue.ceda.ac.uk/uuid/c26a65020a5e4b80b20018f148556681, and https://www.metoffice.gov.uk/hadobs/hadcrut5/, respectively). CMIP6 (Eyring et al. 2016) model data are freely available through the Earth System Grid (see https://esgf-node.llnl.gov/projects/cmip6/). The detection and attribution analysis code used in this study is based on ESMValTool (see https://github.com/ESMValGroup/ESMValTool/blob/gillett20/esmvaltool/; Kirchmeier-Young et al. 2017; Gillett et al. 2021). The code for calculating the time series and reproducing the figures can be found at https://doi.org/10.5281/zenodo.10381597.
REFERENCES
Albrecht, B. A., 1989: Aerosols, cloud microphysics, and fractional cloudiness. Science, 245, 1227–1230, https://doi.org/10.1126/science.245.4923.1227.
Allan, R. P., and Coauthors, 2020: Advances in understanding large-scale responses of the water cycle to climate change. Ann. N. Y. Acad. Sci., 1472, 49–75, https://doi.org/10.1111/nyas.14337.
Allen, M. R., and S. F. B. Tett, 1999: Checking for model consistency in optimal fingerprinting. Climate Dyn., 15, 419–434, https://doi.org/10.1007/s003820050291.
Allen, M. R., and W. J. Ingram, 2002: Constraints on future changes in climate and the hydrologic cycle. Nature, 419, 224–232, https://doi.org/10.1038/nature01092.
Allen, M. R., and P. A. Stott, 2003: Estimating signal amplitudes in optimal fingerprinting, part I: Theory. Climate Dyn., 21, 477–491, https://doi.org/10.1007/s00382-003-0313-9.
Andreae, M. O., C. D. Jones, and P. M. Cox, 2005: Strong present-day aerosol cooling implies a hot future. Nature, 435, 1187–1190, https://doi.org/10.1038/nature03671.
Bellouin, N., and Coauthors, 2020: Bounding global aerosol radiative forcing of climate change. Rev. Geophys., 58, e2019RG000660, https://doi.org/10.1029/2019RG000660.
Bindoff, N. L., and Coauthors, 2013: Detection and attribution of climate change: From global to regional. Climate Change 2013: The Physical Science Basis, T. F. Stocker et al., Eds., Cambridge University Press, 867–952.
Bollasina, M. A., Y. Ming, and V. Ramaswamy, 2011: Anthropogenic aerosols and the weakening of the South Asian summer monsoon. Science, 334, 502–505, https://doi.org/10.1126/science.1204994.
Bonfils, C. J., B. D. Santer, J. C. Fyfe, K. Marvel, T. J. Phillips, and S. R. H. Zimmerman, 2020: Human influence on joint changes in temperature, rainfall and continental aridity. Nat. Climate Change, 10, 726–731, https://doi.org/10.1038/s41558-020-0821-1.
Bony, S., G. Bellon, D. Klocke, S. Sherwood, S. Fermepin, and S. Denvil, 2013: Robust direct effect of carbon dioxide on tropical circulation and regional precipitation. Nat. Geosci., 6, 447–451, https://doi.org/10.1038/ngeo1799.
Boucher, O., and Coauthors, 2013: Clouds and aerosols. Climate Change 2013: The Physical Science Basis, T. F. Stocker et al., Eds., Cambridge University Press, 571–658.
Boucher, O., and Coauthors, 2020: Presentation and evaluation of the IPSL-CM6A-LR climate model. J. Adv. Model. Earth Syst., 12, e2019MS002010, https://doi.org/10.1029/2019MS002010.
Chylek, P., T. J. Vogelsang, J. D. Klett, N. Hengartner, D. Higdon, G. Lesins, and M. K. Dubey, 2016: Indirect aerosol effect increases CMIP5 models’ projected Arctic warming. J. Climate, 29, 1417–1428, https://doi.org/10.1175/JCLI-D-15-0362.1.
Cole, J. N. S., and Coauthors, 2023: The Canadian Atmospheric Model version 5 (CanAM5.0.3). Geosci. Model Dev., 16, 5427–5448, https://doi.org/10.5194/gmd-16-5427-2023.
DelSole, T., L. Trenary, X. Yan, and M. K. Tippett, 2019: Confidence intervals in optimal fingerprinting. Climate Dyn., 52, 4111–4126, https://doi.org/10.1007/s00382-018-4356-3.
Ekman, A. M. L., 2014: Do sophisticated parameterizations of aerosol-cloud interactions in CMIP5 models improve the representation of recent observed temperature trends? J. Geophys. Res. Atmos., 119, 817–832, https://doi.org/10.1002/2013JD020511.
Eyring, V., S. Bony, G. A. Meehl, C. A. Senior, B. Stevens, R. J. Stouffer, and K. E. Taylor, 2016: Overview of the Coupled Model Intercomparison Project Phase 6 (CMIP6) experimental design and organization. Geosci. Model Dev., 9, 1937–1958, https://doi.org/10.5194/gmd-9-1937-2016.
Fläschner, D., T. Mauritsen, and B. Stevens, 2016: Understanding the intermodel spread in global-mean hydrological sensitivity. J. Climate, 29, 801–817, https://doi.org/10.1175/JCLI-D-15-0351.1.
Foote, E., 1856: ART. XXXI.–Circumstances affecting the heat of the sun’s rays. Amer. J. Sci. Arts, 22, 382.
Forster, P., and Coauthors, 2021: The Earth’s energy budget, climate feedbacks, and climate sensitivity. Climate Change 2021: The Physical Science Basis, V. Masson-Delmotte et al., Eds., Cambridge University Press, 923–1054, https://doi.org/10.1017/9781009157896.009.
Forster, P. M., T. Andrews, P. Good, J. M. Gregory, L. S. Jackson, and M. Zelinka, 2013: Evaluating adjusted forcing and model spread for historical and future scenarios in the CMIP5 generation of climate models. J. Geophys. Res. Atmos., 118, 1139–1150, https://doi.org/10.1002/jgrd.50174.
Friedman, A. R., Y.-T. Hwang, J. C. H. Chiang, and D. M. W. Frierson, 2013: Interhemispheric temperature asymmetry over the twentieth century and in future projections. J. Climate, 26, 5419–5433, https://doi.org/10.1175/JCLI-D-12-00525.1.
Friedman, A. R., G. C. Hegerl, A. P. Schurer, S.-Y. Lee, W. Kong, W. Cheng, and J. C. H. Chiang, 2020: Forced and unforced decadal behavior of the interhemispheric SST contrast during the instrumental period (1881–2012): Contextualizing the late 1960s–early 1970s shift. J. Climate, 33, 3487–3509, https://doi.org/10.1175/JCLI-D-19-0102.1.
Gettelman, A., and Coauthors, 2019: High climate sensitivity in the Community Earth System Model version 2 (CESM2). Geophys. Res. Lett., 46, 8329–8337, https://doi.org/10.1029/2019GL083978.
Gillett, N. P., V. K. Arora, D. Matthews, and M. R. Allen, 2013: Constraining the ratio of global warming to cumulative CO2 emissions using CMIP5 simulations. J. Climate, 26, 6844–6858, https://doi.org/10.1175/JCLI-D-12-00476.1.
Gillett, N. P., and Coauthors, 2016: The Detection and Attribution Model Intercomparison Project (DAMIP v1.0) contribution to CMIP6. Geosci. Model Dev., 9, 3685–3697, https://doi.org/10.5194/gmd-9-3685-2016.
Gillett, N. P., and Coauthors, 2021: Constraining human contributions to observed warming since the pre-industrial period. Nat. Climate Change, 11, 207–212, https://doi.org/10.1038/s41558-020-00965-9.
Harris, I., T. J. Osborn, P. Jones, and D. Lister, 2020: Version 4 of the CRU TS monthly high-resolution gridded multivariate climate dataset. Sci. Data, 7, 109, https://doi.org/10.1038/s41597-020-0453-3.
Hasselmann, K., 1993: Optimal fingerprints for the detection of time-dependent climate change. J. Climate, 6, 1957–1971, https://doi.org/10.1175/1520-0442(1993)006<1957:OFFTDO>2.0.CO;2.
Hegerl, G., and F. Zwiers, 2011: Use of models in detection and attribution of climate change. Wiley Interdiscip. Rev.: Climate Change, 2, 570–591, https://doi.org/10.1002/wcc.121.
Hegerl, G. C., K. Hasselmann, U. Cubasch, J. F. B. Mitchell, E. Roeckner, R. Voss, and J. Waszkewitz, 1997: Multi-fingerprint detection and attribution analysis of greenhouse gas, greenhouse gas-plus-aerosol and solar forced climate change. Climate Dyn., 13, 613–634, https://doi.org/10.1007/s003820050186.
Hegerl, G. C., and Coauthors, 2015: Challenges in quantifying changes in the global water cycle. Bull. Amer. Meteor. Soc., 96, 1097–1115, https://doi.org/10.1175/BAMS-D-13-00212.1.
Held, I. M., and B. J. Soden, 2006: Robust responses of the hydrological cycle to global warming. J. Climate, 19, 5686–5699, https://doi.org/10.1175/JCLI3990.1.
Hoesly, R. M., and Coauthors, 2018: Historical (1750–2014) anthropogenic emissions of reactive gases and aerosols from the Community Emissions Data System (CEDS). Geosci. Model Dev., 11, 369–408, https://doi.org/10.5194/gmd-11-369-2018.
Huffman, G. J., R. F. Adler, A. Behrangi, D. T. Bolvin, E. J. Nelkin, G. Gu, and M. R. Ehsani, 2023: The new version 3.2 Global Precipitation Climatology Project (GPCP) monthly and daily precipitation products. J. Climate, 36, 7635–7655, https://doi.org/10.1175/JCLI-D-23-0123.1.
IPCC, 2021a: Climate Change 2021: The Physical Science Basis. Cambridge University Press, 2391 pp., https://doi.org/10.1017/9781009157896.
IPCC, 2021b: Summary for policymakers. Climate Change 2021: The Physical Science Basis, V. Masson-Delmotte et al., Eds., Cambridge University Press, 3–32, https://doi.org/10.1017/9781009157896.001.
Jones, G. S., P. A. Stott, and N. Christidis, 2013: Attribution of observed historical near–surface temperature variations to anthropogenic and natural causes using CMIP5 simulations. J. Geophys. Res. Atmos., 118, 4001–4024, https://doi.org/10.1002/jgrd.50239.
Jones, G. S., P. A. Stott, and J. F. B. Mitchell, 2016: Uncertainties in the attribution of greenhouse gas warming and implications for climate prediction. J. Geophys. Res. Atmos., 121, 6969–6992, https://doi.org/10.1002/2015JD024337.
Jones, P. W., 1999: First- and second-order conservative remapping schemes for grids in spherical coordinates. Mon. Wea. Rev., 127, 2204–2210, https://doi.org/10.1175/1520-0493(1999)127<2204:FASOCR>2.0.CO;2.
Kirchmeier-Young, M. C., F. W. Zwiers, and N. P. Gillett, 2017: Attribution of extreme events in Arctic sea ice extent. J. Climate, 30, 553–571, https://doi.org/10.1175/JCLI-D-16-0412.1.
Larson, E. J. L., and R. W. Portmann, 2019: Anthropogenic aerosol drives uncertainty in future climate mitigation efforts. Sci. Rep., 9, 16538, https://doi.org/10.1038/s41598-019-52901-3.
Li, Y., K. Chen, J. Yan, and X. Zhang, 2021: Uncertainty in optimal fingerprinting is underestimated. Environ. Res. Lett., 16, 084043, https://doi.org/10.1088/1748-9326/ac14ee.
Lorenz, C., and H. Kunstmann, 2012: The hydrological cycle in three state-of-the-art reanalyses: Intercomparison and performance analysis. J. Hydrometeor., 13, 1397–1420, https://doi.org/10.1175/JHM-D-11-088.1.
Lund, M. T., G. Myhre, and B. H. Samset, 2019: Anthropogenic aerosol forcing under the shared socioeconomic pathways. Atmos. Chem. Phys., 19, 13 827–13 839, https://doi.org/10.5194/acp-19-13827-2019.
Lurton, T., and Coauthors, 2020: Implementation of the CMIP6 forcing data in the IPSL-CM6A-LR Model. J. Adv. Model. Earth Syst., 12, e2019MS001940, https://doi.org/10.1029/2019MS001940.
Ma, S., T. Wang, J. Yan, and X. Zhang, 2023: Optimal fingerprinting with estimating equations. J. Climate, 36, 7109–7122, https://doi.org/10.1175/JCLI-D-22-0681.1.
McDuffie, E. E., and Coauthors, 2020: A global anthropogenic emission inventory of atmospheric pollutants from sector- and fuel-specific sources (1970–2017): An application of the Community Emissions Data System (CEDS). Earth Syst. Sci. Data, 12, 3413–3442, https://doi.org/10.5194/essd-12-3413-2020.
Meehl, G. A., C. A. Senior, V. Eyring, G. Flato, J.-F. Lamarque, R. J. Stouffer, K. E. Taylor, and M. Schlund, 2020: Context for interpreting equilibrium climate sensitivity and transient climate response from the CMIP6 Earth System Models. Sci. Adv., 6, eaba1981, https://doi.org/10.1126/sciadv.aba1981.
Ming, Y., and V. Ramaswamy, 2009: Nonlinear climate and hydrological responses to aerosol effects. J. Climate, 22, 1329–1339, https://doi.org/10.1175/2008JCLI2362.1.
Morice, C. P., and Coauthors, 2021: An updated assessment of near-surface temperature change from 1850: The HadCRUT5 data set. J. Geophys. Res. Atmos., 126, e2019JD032361, https://doi.org/10.1029/2019JD032361.
Norris, J., A. Hall, C. W. Thackeray, D. Chen, and G. D. Madakumbura, 2022: Evaluating hydrologic sensitivity in CMIP6 models: Anthropogenic forcing versus ENSO. J. Climate, 35, 3349–3362, https://doi.org/10.1175/JCLI-D-21-0842.1.
Parsons, L. A., M. K. Brennan, R. C. J. Wills, and C. Proistosescu, 2020: Magnitudes and spatial patterns of interdecadal temperature variability in CMIP6. Geophys. Res. Lett., 47, e2019GL086588, https://doi.org/10.1029/2019GL086588.
Pendergrass, A. G., 2020: The global-mean precipitation response to CO2-induced warming in CMIP6 models. Geophys. Res. Lett., 47, e2020GL089964, https://doi.org/10.1029/2020GL089964.
Persad, G. G., B. H. Samset, and L. J. Wilcox, 2022: Aerosols must be included in climate risk assessments. Nature, 611, 662–664, https://doi.org/10.1038/d41586-022-03763-9.
Polson, D., and G. C. Hegerl, 2017: Strengthening contrast between precipitation in tropical wet and dry regions. Geophys. Res. Lett., 44, 365–373, https://doi.org/10.1002/2016GL071194.
Polson, D., G. C. Hegerl, R. P. Allan, and B. B. Sarojini, 2013a: Have greenhouse gases intensified the contrast between wet and dry regions? Geophys. Res. Lett., 40, 4783–4787, https://doi.org/10.1002/grl.50923.
Polson, D., G. C. Hegerl, X. Zhang, and T. J. Osborn, 2013b: Causes of robust seasonal land precipitation changes. J. Climate, 26, 6679–6697, https://doi.org/10.1175/JCLI-D-12-00474.1.
Polson, D., M. Bollasina, G. C. Hegerl, and L. J. Wilcox, 2014: Decreased monsoon precipitation in the Northern Hemisphere due to anthropogenic aerosols. Geophys. Res. Lett., 41, 6023–6029, https://doi.org/10.1002/2014GL060811.
Ramanathan, V., P. J. Crutzen, J. T. Kiehl, and D. Rosenfeld, 2001: Aerosols, climate, and the hydrological cycle. Science, 294, 2119–2124, https://doi.org/10.1126/science.1064034.
Randles, C. A., P. R. Colarco, and A. da Silva, 2013: Direct and semi-direct aerosol effects in the NASA GEOS-5 AGCM: Aerosol-climate interactions due to prognostic versus prescribed aerosols. J. Geophys. Res. Atmos., 118, 149–169, https://doi.org/10.1029/2012JD018388.
Ribes, A., and L. Terray, 2013: Application of regularised optimal fingerprinting to attribution. Part II: Application to global near-surface temperature. Climate Dyn., 41, 2837–2853, https://doi.org/10.1007/s00382-013-1736-6.
Ribes, A., S. Planton, and L. Terray, 2013: Application of regularised optimal fingerprinting to attribution. Part I: Method, properties and idealised analysis. Climate Dyn., 41, 2817–2836, https://doi.org/10.1007/s00382-013-1735-7.
Ribes, A., N. P. Gillett, and F. W. Zwiers, 2015: Designing detection and attribution simulations for CMIP6 to optimize the estimation of greenhouse gas–induced warming. J. Climate, 28, 3435–3438, https://doi.org/10.1175/JCLI-D-14-00691.1.
Roehrig, R., and Coauthors, 2020: The CNRM global atmosphere model ARPEGE-Climat 6.3: Description and evaluation. J. Adv. Model. Earth Syst., 12, e2020MS002075, https://doi.org/10.1029/2020MS002075.
Schneider, T., T. Bischoff, and G. H. Haug, 2014: Migrations and dynamics of the intertropical convergence zone. Nature, 513, 45–53, https://doi.org/10.1038/nature13636.
Schurer, A., G. Hegerl, A. Ribes, D. Polson, C. Morice, and S. Tett, 2018: Estimating the transient climate response from observed warming. J. Climate, 31, 8645–8663, https://doi.org/10.1175/JCLI-D-17-0717.1.
Schurer, A. P., A. P. Ballinger, A. R. Friedman, and G. C. Hegerl, 2020: Human influence strengthens the contrast between tropical wet and dry regions. Environ. Res. Lett., 15, 104026, https://doi.org/10.1088/1748-9326/ab83ab.
Shindell, D. T., G. Faluvegi, L. Rotstayn, and G. Milly, 2015: Spatial patterns of radiative forcing and surface temperature response. J. Geophys. Res. Atmos., 120, 5385–5403, https://doi.org/10.1002/2014JD022752.
Shiogama, H., D. A. Stone, T. Nagashima, T. Nozawa, and S. Emori, 2013: On the linear additivity of climate forcing-response relationships at global and continental scales. Int. J. Climatol., 33, 2542–2550, https://doi.org/10.1002/joc.3607.
Smith, C. J., and Coauthors, 2020: Effective radiative forcing and adjustments in CMIP6 models. Atmos. Chem. Phys., 20, 9591–9618, https://doi.org/10.5194/acp-20-9591-2020.
Stjern, C. W., B. H. Samset, O. Boucher, T. Iversen, J.-F. Lamarque, G. Myhre, D. Shindell, and T. Takemura, 2020: How aerosols and greenhouse gases influence the diurnal temperature range. Atmos. Chem. Phys., 20, 13 467–13 480, https://doi.org/10.5194/acp-20-13467-2020.
Swart, N. C., and Coauthors, 2019: The Canadian Earth System Model version 5 (CanESM5.0.3). Geosci. Model Dev., 12, 4823–4873, https://doi.org/10.5194/gmd-12-4823-2019.
Tatebe, H., and Coauthors, 2019: Description and basic evaluation of simulated mean state, internal variability, and climate sensitivity in MIROC6. Geosci. Model Dev., 12, 2727–2765, https://doi.org/10.5194/gmd-12-2727-2019.
Twomey, S., 1977: The influence of pollution on the shortwave albedo of clouds. J. Atmos. Sci., 34, 1149–1152, https://doi.org/10.1175/1520-0469(1977)034<1149:TIOPOT>2.0.CO;2.
Undorf, S., M. A. Bollasina, and G. C. Hegerl, 2018: Impacts of the 1900–74 increase in anthropogenic aerosol emissions from North America and Europe on Eurasian summer climate. J. Climate, 31, 8381–8399, https://doi.org/10.1175/JCLI-D-17-0850.1.
Van Huffel, S., and J. Vandewalle, 1991: Basic principles of the total least squares problem. The Total Least Squares Problem: Computational Aspects and Analysis, SIAM, 27–47.
Voldoire, A., and Coauthors, 2019: Evaluation of CMIP6 deck experiments with CNRM-CM6-1. J. Adv. Model. Earth Syst., 11, 2177–2213, https://doi.org/10.1029/2019MS001683.
Walters, D., and Coauthors, 2019: The Met Office Unified Model global atmosphere 7.0/7.1 and JULES global land 7.0 configurations. Geosci. Model Dev., 12, 1909–1963, https://doi.org/10.5194/gmd-12-1909-2019.
Wang, C., B. J. Soden, W. Yang, and G. A. Vecchi, 2021: Compensation between cloud feedback and aerosol-cloud interaction in CMIP6 models. Geophys. Res. Lett., 48, e2020GL091024, https://doi.org/10.1029/2020GL091024.
Wilcox, L. J., E. J. Highwood, and N. J. Dunstone, 2013: The influence of anthropogenic aerosol on multi-decadal variations of historical global climate. Environ. Res. Lett., 8, 024033, https://doi.org/10.1088/1748-9326/8/2/024033.
Wilcox, L. J., E. J. Highwood, B. B. B. Booth, and K. S. Carslaw, 2015: Quantifying sources of inter-model diversity in the cloud albedo effect. Geophys. Res. Lett., 42, 1568–1575, https://doi.org/10.1002/2015GL063301.
Wild, M., A. Ohmura, and K. Makowski, 2007: Impact of global dimming and brightening on global warming. Geophys. Res. Lett., 34, L04702, https://doi.org/10.1029/2006GL028031.
Williams, K. D., and Coauthors, 2018: The Met Office global coupled model 3.0 and 3.1 (GC3.0 and GC3.1) configurations. J. Adv. Model. Earth Syst., 10, 357–380, https://doi.org/10.1002/2017MS001115.
Yukimoto, S., and Coauthors, 2019: The Meteorological Research Institute Earth System Model version 2.0, MRI-ESM2.0: Description and basic evaluation of the physical component. J. Meteor. Soc. Japan, 97, 931–965, https://doi.org/10.2151/jmsj.2019-051.
Zelinka, M. D., T. Andrews, P. M. Forster, and K. E. Taylor, 2014: Quantifying components of aerosol-cloud-radiation interactions in climate models. J. Geophys. Res. Atmos., 119, 7599–7615, https://doi.org/10.1002/2014JD021710.
Zelinka, M. D., T. A. Myers, D. T. McCoy, S. Po-Chedley, P. M. Caldwell, P. Ceppi, S. A. Klein, and K. E. Taylor, 2020: Causes of higher climate sensitivity in CMIP6 models. Geophys. Res. Lett., 47, e2019GL085782, https://doi.org/10.1029/2019GL085782.
Zhang, S., P. Stier, and D. Watson-Parris, 2021: On the contribution of fast and slow responses to precipitation changes caused by aerosol perturbations. Atmos. Chem. Phys., 21, 10 179–10 197, https://doi.org/10.5194/acp-21-10179-2021.
Zhang, X., F. W. Zwiers, G. C. Hegerl, F. H. Lambert, N. P. Gillett, S. Solomon, P. A. Stott, and T. Nozawa, 2007: Detection of human influence on twentieth-century precipitation trends. Nature, 448, 461–465, https://doi.org/10.1038/nature06025.