Misplaced Pages

Hockey stick graph (global temperature)

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.

A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency ") in mathematics , especially in statistics . Each attempts to summarize or typify a given group of data , illustrating the magnitude and sign of the data set . Which of these measures is most illuminating depends on what is being measured, and on context and purpose.

#346653

159-448: Hockey stick graphs present the global or hemispherical mean temperature record of the past 500 to 2000 years as shown by quantitative climate reconstructions based on climate proxy records. These reconstructions have consistently shown a slow long term cooling trend changing into relatively rapid warming in the 20th century, with the instrumental temperature record by 2000 exceeding earlier temperatures. The term hockey stick graph

318-537: A Medieval Warm Period from around 900 to 1300, followed by Little Ice Age . This was the basis of a "schematic diagram" featured in the IPCC First Assessment Report of 1990 beside cautions that the medieval warming might not have been global. The use of indicators to get quantitative estimates of the temperature record of past centuries was developed, and by the late 1990s a number of competing teams of climatologists found indications that recent warming

477-425: A color wheel —there is no mean to the set of all colors. In these situations, you must decide which mean is most useful. You can do this by adjusting the values before averaging, or by using a specialized approach for the mean of circular quantities . The Fréchet mean gives a manner for determining the "center" of a mass distribution on a surface or, more generally, Riemannian manifold . Unlike many other means,

636-434: A p -dimensional ellipsoid to the data, where each axis of the ellipsoid represents a principal component. If some axis of the ellipsoid is small, then the variance along that axis is also small. To find the axes of the ellipsoid, we must first center the values of each variable in the dataset on 0 by subtracting the mean of the variable's observed values from each of those values. These transformed values are used instead of

795-594: A principal component analysis step to summarise these proxy networks, but from 2001 Mann stopped using this method and introduced a multivariate Climate Field Reconstruction (CFR) technique based on the RegEM method which did not require this PCA step. In May 2002 Mann and Scott Rutherford published a paper on testing methods of climate reconstruction which discussed this technique. By adding artificial noise to actual temperature records or to model simulations they produced synthetic datasets which they called "pseudoproxies". When

954-479: A random variable having that distribution. If the random variable is denoted by X {\displaystyle X} , then the mean is also known as the expected value of X {\displaystyle X} (denoted E ( X ) {\displaystyle E(X)} ). For a discrete probability distribution , the mean is given by ∑ x P ( x ) {\displaystyle \textstyle \sum xP(x)} , where

1113-400: A truncated mean . It involves discarding given parts of the data at the top or the bottom end, typically an equal amount at each end and then taking the arithmetic mean of the remaining data. The number of values removed is indicated as a percentage of the total number of values. The interquartile mean is a specific example of a truncated mean. It is simply the arithmetic mean after removing

1272-493: A 1991 study and a 1991 book showing methodology and examples of how to produces maps showing climate developments in North America over time. These methods had been used for regional reconstructions of temperatures, and other aspects such as rainfall. As part of his PhD research, Michael E. Mann worked with seismologist Jeffrey Park on developing statistical techniques to find long term oscillations of natural variability in

1431-827: A co-author on the paper, had confirmed agreement with these points, and a later paper by Cook, Esper and D'Arrigo reconsidered the earlier paper's conclusions along these lines. Lonnie Thompson published a paper on "Tropical Glacier and Ice Core Evidence of Climate Change" in January 2003, featuring Figure 7 showing graphs based on ice cores closely resembling a graph based on the MBH99 reconstruction, combined with thermometer readings from Jones et al. 1999. In March 2001 Tapio Schneider published his regularized expectation–maximization (RegEM) technique for analysis of incomplete climate data. The original MBH98 and MBH99 papers avoided undue representation of large numbers of tree ring proxies by using

1590-458: A comparison with the thermometer record to check that recent proxy data were valid. Jones thought the study would provide important comparisons with the findings of climate modeling , which showed a "pretty reasonable" fit to proxy evidence. A commentary on MBH98 by Jones was published in Science on 24 April 1998. He noted that it used almost all the available long term proxy climate series, "and if

1749-399: A data vector x ( i ) can therefore be given as a score t k ( i ) = x ( i ) ⋅ w ( k ) in the transformed coordinates, or as the corresponding vector in the space of the original variables, { x ( i ) ⋅ w ( k ) } w ( k ) , where w ( k ) is the k th eigenvector of X X . The full principal components decomposition of X can therefore be given as where W

SECTION 10

#1732870051347

1908-427: A data vector x ( i ) from an original space of p variables to a new space of p variables which are uncorrelated over the dataset. However, not all the principal components need to be kept. Keeping only the first L principal components, produced by using only the first L eigenvectors, gives the truncated transformation where the matrix T L now has n rows but only L columns. In other words, PCA learns

2067-449: A decadal basis rather than showing individual years, and produced a single time series so did not show a spatial pattern of relative temperatures for different regions. The IPCC Second Assessment Report (SAR) of 1996 featured Figure 3.20 showing the Bradley & Jones 1993 decadal summer temperature reconstruction for the northern hemisphere, overlaid with a 50-year smoothed curve and with

2226-487: A discussion of the differences between PCA and factor analysis see Ch. 7 of Jolliffe's Principal Component Analysis ), Eckart–Young theorem (Harman, 1960), or empirical orthogonal functions (EOF) in meteorological science (Lorenz, 1956), empirical eigenfunction decomposition (Sirovich, 1987), quasiharmonic modes (Brooks et al., 1988), spectral decomposition in noise and vibration, and empirical modal analysis in structural dynamics. PCA can be thought of as fitting

2385-463: A focus of dispute for those opposed to the strengthening scientific consensus that late 20th century warmth was exceptional. In 2003, as lobbying over the 1997 Kyoto Protocol intensified, a paper claiming greater medieval warmth was quickly dismissed by scientists in the Soon and Baliunas controversy . Later in 2003, Stephen McIntyre and Ross McKitrick published McIntyre & McKitrick 2003b disputing

2544-417: A function f ( x ) {\displaystyle f(x)} . Intuitively, a mean of a function can be thought of as calculating the area under a section of a curve, and then dividing by the length of that section. This can be done crudely by counting squares on graph paper, or more precisely by integration . The integration formula is written as: In this case, care must be taken to make sure that

2703-552: A future ice age, but after 1976 he supported the emerging view that greenhouse gas emissions caused by humanity would cause detectable global warming "by about A.D. 2000". The first quantitative reconstruction of Northern Hemisphere (NH) annual mean temperatures was published in 1979 by Brian Groveman and Helmut Landsberg. They used "a short-cut method" based on their earlier paper which showed that 9 instrumental stations could adequately represent an extensive gridded instrumental series, and reconstructed temperatures from 1579 to 1880 on

2862-419: A high-resolution global reconstruction. To relate this data to measured temperatures, they used principal component analysis (PCA) to find the leading patterns, or principal components, of instrumental temperature records during the calibration period from 1902 to 1980. Their method was based on separate multiple regressions between each proxy record (or summary) and all of the leading principal components of

3021-403: A linear transformation t = W L T x , x ∈ R p , t ∈ R L , {\displaystyle t=W_{L}^{\mathsf {T}}x,x\in \mathbb {R} ^{p},t\in \mathbb {R} ^{L},} where the columns of p × L matrix W L {\displaystyle W_{L}} form an orthogonal basis for

3180-454: A new vector of principal component scores t ( i ) = ( t 1 , … , t l ) ( i ) {\displaystyle \mathbf {t} _{(i)}=(t_{1},\dots ,t_{l})_{(i)}} , given by in such a way that the individual variables t 1 , … , t l {\displaystyle t_{1},\dots ,t_{l}} of t considered over

3339-430: A sample x 1 , x 2 , … , x n {\displaystyle x_{1},x_{2},\ldots ,x_{n}} , usually denoted by x ¯ {\displaystyle {\bar {x}}} , is the sum of the sampled values divided by the number of items in the sample. For example, the arithmetic mean of five values: 4, 36, 45, 50, 75 is: The geometric mean

SECTION 20

#1732870051347

3498-404: A separate curve plotting instrumental thermometer data from the 1850s onwards. It stated that in this record, warming since the late 19th century was unprecedented. The section proposed that "The data from the last 1000 years are the most useful for determining the scales of natural climate variability". Recent studies including the 1994 reconstruction by Hughes and Diaz questioned how widespread

3657-415: A sequence of p {\displaystyle p} unit vectors , where the i {\displaystyle i} -th vector is the direction of a line that best fits the data while being orthogonal to the first i − 1 {\displaystyle i-1} vectors. Here, a best-fitting line is defined as one that minimizes the average squared perpendicular distance from

3816-431: A simplified figure for the cover of the short annual World Meteorological Organization report, which lacks the status of the more important IPCC reports. Two fifty-year smoothed curves going back to 1000 were shown, from MBH99 and Jones et al. (1998), with a third curve to 1400 from Briffa's new paper, combined with modern temperature data bringing the lines up to 1999: in 2010 the lack of a clarity about this change of data

3975-547: A study of 1,000 years of tree ring data from Tasmania which, like similar studies, did not allow for possible overestimate of warming due to increased CO 2 levels having a fertilisation effect on tree growth. It noted the suggestion of Bradley et al. 1991 that instrumental records in specific areas could be combined with paleoclimate data for increased detail back to the 18th century. Archives of climate proxies were developed: in 1993 Raymond S. Bradley and Phil Jones composited historical records, tree-rings and ice cores for

4134-474: A temperature drop of almost 0.5 °C during the Little Ice Age , and increased solar output might explain the rise in early 20th century temperatures. A reconstruction of Arctic temperatures over four centuries by Overpeck et al. 1997 reached similar conclusions, but both these studies came up against the limitations of the climate reconstructions at that time which only resolved temperature fluctuations on

4293-399: A two-dimensional diagram; whereas if two directions through the data (or two of the original variables) are chosen at random, the clusters may be much less spread apart from each other, and may in fact be much more likely to substantially overlay each other, making them indistinguishable. Similarly, in regression analysis , the larger the number of explanatory variables allowed, the greater

4452-467: A very useful step for visualising and processing high-dimensional datasets, while still retaining as much of the variance in the dataset as possible. For example, selecting L  = 2 and keeping only the first two principal components finds the two-dimensional plane through the high-dimensional dataset in which the data is most spread out, so if the data contains clusters these too may be most spread out, and therefore most visible to be plotted out in

4611-412: Is a linear dimensionality reduction technique with applications in exploratory data analysis , visualization and data preprocessing . The data is linearly transformed onto a new coordinate system such that the directions (principal components) capturing the largest variation in the data can be easily identified. The principal components of a collection of points in a real coordinate space are

4770-535: Is a p -by- p matrix of weights whose columns are the eigenvectors of X X . The transpose of W is sometimes called the whitening or sphering transformation . Columns of W multiplied by the square root of corresponding eigenvalues, that is, eigenvectors scaled up by the variances, are called loadings in PCA or in Factor analysis. X X itself can be recognized as proportional to the empirical sample covariance matrix of

4929-404: Is a certain amount of skill. We can actually say something, although there are large uncertainties." In considering the 1998 Jones et al. reconstruction which went back a thousand years, Mann, Bradley and Hughes reviewed their own research and reexamined 24 proxy records which extended back before 1400. Mann carried out a series of statistical sensitivity tests , removing each proxy in turn to see

Hockey stick graph (global temperature) - Misplaced Pages Continue

5088-401: Is an average that is useful for sets of positive numbers, that are interpreted according to their product (as is the case with rates of growth) and not their sum (as is the case with the arithmetic mean): For example, the geometric mean of five values: 4, 36, 45, 50, 75 is: The harmonic mean is an average which is useful for sets of numbers which are defined in relation to some unit , as in

5247-422: Is below and half is above. The mode income is the most likely income and favors the larger number of people with lower incomes. While the median and mode are often more intuitive measures for such skewed data, many skewed distributions are in fact best described by their mean, including the exponential and Poisson distributions. The mean of a probability distribution is the long-run arithmetic average value of

5406-636: Is quite a bit of work to be done in reducing these uncertainties." Climatologist Tom Wigley welcomed the progress made in the study, but doubted if proxy data could ever be wholly convincing in detecting the human contribution to changing climate. Phil Jones of the UEA Climatic Research Unit told the New York Times he was doubtful about adding the 150-year thermometer record to extend the proxy reconstruction, and compared this with putting together apples and oranges; Mann et al. said they used

5565-430: Is sensitive to the scaling of the variables. If we have just two variables and they have the same sample variance and are completely correlated, then the PCA will entail a rotation by 45° and the "weights" (they are the cosines of rotation) for the two variables with respect to the principal component will be equal. But if we multiply all values of the first variable by 100, then the first principal component will be almost

5724-487: Is the probability density function . In all cases, including those in which the distribution is neither discrete nor continuous, the mean is the Lebesgue integral of the random variable with respect to its probability measure . The mean need not exist or be finite; for some probability distributions the mean is infinite ( +∞ or −∞ ), while for others the mean is undefined . The generalized mean , also known as

5883-413: Is the arithmetic average of the values; however, for skewed distributions , the mean is not necessarily the same as the middle value (median), or the most likely value (mode). For example, mean income is typically skewed upwards by a small number of people with very large incomes, so that the majority have an income lower than the mean. By contrast, the median income is the level at which half the population

6042-412: Is the chance of overfitting the model, producing conclusions that fail to generalise to other datasets. One approach, especially when there are strong correlations between different possible explanatory variables, is to reduce them to a few principal components and then run the regression against them, a method called principal component regression . Dimensionality reduction may also be appropriate when

6201-419: Is the square diagonal matrix with the singular values of X and the excess zeros chopped off that satisfies Σ ^ 2 = Σ T Σ {\displaystyle \mathbf {{\hat {\Sigma }}^{2}} =\mathbf {\Sigma } ^{\mathsf {T}}\mathbf {\Sigma } } . Comparison with the eigenvector factorization of X X establishes that

6360-480: Is the sum of the values divided by the number of values. The arithmetic mean of a set of numbers x 1 , x 2 , ..., x n is typically denoted using an overhead bar , x ¯ {\displaystyle {\bar {x}}} . If the numbers are from observing a sample of a larger group , the arithmetic mean is termed the sample mean ( x ¯ {\displaystyle {\bar {x}}} ) to distinguish it from

6519-521: The IPCC Third Assessment Report (TAR) included a subsection on multi-proxy synthesis of recent temperature change. This noted five earlier large-scale palaeoclimate reconstructions, then discussed the Mann, Bradley & Hughes 1998 reconstruction going back to 1400 AD and its extension back to 1000 AD in Mann, Bradley & Hughes 1999 (MBH99), while emphasising the substantial uncertainties in

Hockey stick graph (global temperature) - Misplaced Pages Continue

6678-779: The L features (the components of representation t ) that are decorrelated. By construction, of all the transformed data matrices with only L columns, this score matrix maximises the variance in the original data that has been preserved, while minimising the total squared reconstruction error ‖ T W T − T L W L T ‖ 2 2 {\displaystyle \|\mathbf {T} \mathbf {W} ^{T}-\mathbf {T} _{L}\mathbf {W} _{L}^{T}\|_{2}^{2}} or ‖ X − X L ‖ 2 2 {\displaystyle \|\mathbf {X} -\mathbf {X} _{L}\|_{2}^{2}} . Such dimensionality reduction can be

6837-498: The National Oceanic and Atmospheric Administration 's Geophysical Fluid Dynamics Laboratory , Jerry Mahlman nicknamed the graph the "hockey stick", with the slow cooling trend the "stick", and the anomalous 20th century warming the "blade". Briffa and Tim Osborn critically examined MBH99 in a May 1999 detailed study of the uncertainties of various proxies. They raised questions later adopted by critics of Mann's work, including

6996-676: The Synthesis Report - Questions . The Working Group 1 scientific basis report was agreed unanimously by all member government representatives in January 2001 at a meeting held in Shanghai , China. A large poster of the IPCC illustration based on the MBH99 graph formed the backdrop when Sir John T. Houghton , as co-chair of the working group, presented the report in an announcement shown on television, leading to wide publicity. The Huang, Pollack & Shen 2000 borehole temperature reconstruction covering

7155-400: The arithmetic mean (AM), the geometric mean (GM), and the harmonic mean (HM). These means were studied with proportions by Pythagoreans and later generations of Greek mathematicians because of their importance in geometry and music. The arithmetic mean (or simply mean or average ) of a list of numbers, is the sum of all of the numbers divided by their count. Similarly, the mean of

7314-420: The cross-covariance between two datasets while PCA defines a new orthogonal coordinate system that optimally describes variance in a single dataset. Robust and L1-norm -based variants of standard PCA have also been proposed. PCA was invented in 1901 by Karl Pearson , as an analogue of the principal axis theorem in mechanics; it was later independently developed and named by Harold Hotelling in

7473-441: The group mean (or expected value ) of the underlying distribution, denoted μ {\displaystyle \mu } or μ x {\displaystyle \mu _{x}} . Outside probability and statistics, a wide range of other notions of mean are often used in geometry and mathematical analysis ; examples are given below. In mathematics, the three classical Pythagorean means are

7632-530: The instrumental temperature record of global surface temperatures over the last 140 years; Mann & Park 1993 showed patterns relating to the El Niño–Southern Oscillation , and Mann & Park 1994 found what was later termed the Atlantic multidecadal oscillation . They then teamed up with Raymond S. Bradley to use these techniques on the dataset from his Bradley & Jones 1993 study with

7791-801: The n rows represents a different repetition of the experiment, and each of the p columns gives a particular kind of feature (say, the results from a particular sensor). Mathematically, the transformation is defined by a set of size l {\displaystyle l} of p -dimensional vectors of weights or coefficients w ( k ) = ( w 1 , … , w p ) ( k ) {\displaystyle \mathbf {w} _{(k)}=(w_{1},\dots ,w_{p})_{(k)}} that map each row vector x ( i ) = ( x 1 , … , x p ) ( i ) {\displaystyle \mathbf {x} _{(i)}=(x_{1},\dots ,x_{p})_{(i)}} of X to

7950-527: The polar regions and the tropics , they used principal component analysis (PCA) to produce PC summaries representing these large datasets, and then treated each summary as a proxy record in their CFR analysis. Networks represented in this way included the North American tree ring network (NOAMER) and Eurasia . The primary aim of CFR methods was to provide the spatially resolved reconstructions essential for coherent geophysical understanding of how parts of

8109-403: The "Composite Plus Scaling" (CPS) method which was subsequently used by most large-scale climate reconstructions of hemispheric or global average temperatures. In this method, also known as "Composite Plus Scale", selected climate proxy records were standardized before being averaged (composited), and then centred and scaled to provide a quantitative estimate of the target temperature series for

SECTION 50

#1732870051347

8268-401: The "Medieval Climatic Optimum" from the "late tenth to early thirteenth centuries (about AD 950-1250)", followed by a cooler period of the Little Ice Age which ended only in the middle to late nineteenth century. The report discussed the difficulties with proxy data, "mainly pollen remains, lake varves and ocean sediments, insect and animal remains, glacier termini" but considered tree ring data

8427-497: The 1930s. Depending on the field of application, it is also named the discrete Karhunen–Loève transform (KLT) in signal processing , the Hotelling transform in multivariate quality control, proper orthogonal decomposition (POD) in mechanical engineering, singular value decomposition (SVD) of X (invented in the last quarter of the 19th century ), eigenvalue decomposition (EVD) of X X in linear algebra, factor analysis (for

8586-408: The 1950s) down to a cooler Little Ice Age before rising sharply in the 20th century. Thermometer data shown with a dotted line overlapped the reconstruction for a calibration period from 1902 to 1980, then continued sharply up to 1998. A shaded area showed uncertainties to two standard error limits, in medieval times rising almost as high as recent temperatures. When Mann gave a talk about the study to

8745-451: The 1960s that accurate use of tree rings as climate proxies for reconstructions was pioneered by Harold C. Fritts . In 1965 Hubert Lamb , a pioneer of historical climatology , generalised from temperature records of central England by using historical, botanical and archeological evidence to popularise the idea of a Medieval Warm Period from around 900 to 1300, followed by a cold epoch culminating between 1550 and 1700. In 1972 he became

8904-513: The 1988 formation of the Intergovernmental Panel on Climate Change to produce reports subject to detailed approval by government delegates. The IPCC First Assessment Report in 1990 noted evidence that Holocene climatic optimum around 5,000-6,000 years ago had been warmer than the present (at least in summer) and that in some areas there had been exceptional warmth during "a shorter Medieval Warm Period (which may not have been global)",

9063-482: The 1990s was the warmest decade and 1998 the warmest year" in the past 1,000 years. Versions of these graphs also featured less prominently in the short Synthesis Report Summary for Policymakers , which included a sentence stating that "The increase in surface temperature over the 20th century for the Northern Hemisphere is likely to have been greater than that for any other century in the last thousand years", and

9222-456: The 19th century physicists John Tyndall and Svante Arrhenius who found the greenhouse gas effect of carbon dioxide (CO 2 ) in the atmosphere to explain how past ice ages had ended. From 1919 to 1923, Alfred Wegener did pioneering work on reconstructing the climate of past eras in collaboration with Milutin Milanković , publishing Die Klimate der geologischen Vorzeit ("The Climates of

9381-444: The 20th century. After this around a decade elapsed before Gordon Jacoby and Rosanne D'Arrigo produced the next quantitative NH reconstruction, published in 1989. This was the first based entirely on non-instrumental records, and used tree rings. They reconstructed northern hemisphere annual temperatures since 1671 on the basis of boreal North American tree ring data from 11 distinct regions. From this, they concluded that recent warming

9540-503: The Asian monsoon region, the El Niño–Southern Oscillation region and the Atlantic region. Areas where more data was needed were to be identified, and there was a need for improved data exchange with computer-based archiving and translation to give researchers access to worldwide paleoclimate information. The IPCC supplementary report, 1992 , reviewed progress on various proxies. These included

9699-501: The Fréchet mean is defined on a space whose elements cannot necessarily be added together or multiplied by scalars. It is sometimes also known as the Karcher mean (named after Hermann Karcher). In geometry, there are thousands of different definitions for the center of a triangle that can all be interpreted as the mean of a triangular set of points in the plane. This is an approximation to

SECTION 60

#1732870051347

9858-458: The Geological Past") together with Wladimir Köppen, in 1924. In the 1930s Guy Stewart Callendar compiled temperature records to look for changes. Wilmot H. Bradley showed that annual varves in lake beds showed climate cycles, and A. E. Douglass found that tree rings could track past climatic changes but these were thought to only show random variations in the local region. It was only in

10017-465: The MBH98 methodology to extend their study back to 1000. A version of the MBH99 graph was featured prominently in the 2001 IPCC Third Assessment Report (TAR), which also drew on Jones et al. 1998 and three other reconstructions to support the conclusion that, in the Northern Hemisphere, the 1990s was likely to have been the warmest decade and 1998 the warmest year during the past 1,000 years. The graph became

10176-598: The McIntyre and McKitrick methodology. Political disputes led to the formation of a panel of scientists convened by the United States National Research Council , their North Report in 2006 supported Mann's findings with some qualifications, including agreeing that there were some statistical failings but these had little effect on the result. More than two dozen reconstructions , using various statistical methods and combinations of proxy records, support

10335-462: The Medieval Warm Period had been at any one time, thus it was not possible "to conclude that global temperatures in the Medieval Warm Period were comparable to the warm decades of the late 20th century." The SAR concluded, "it appears that the 20th century has been at least as warm as any century since at least 1400 AD. In at least some areas, the recent period appears to be warmer than has been

10494-474: The Northern Hemisphere from 1400 up to the 1970s to produce a decadal reconstruction. Like later reconstructions including the MBH "hockey stick" studies, the Bradley & Jones 1993 reconstruction indicated a slow cooling trend followed by an exceptional temperature rise in the 20th century. Their study also used the modern instrumental temperature record to evaluate how well the regions covered by proxies represented

10653-587: The PCA. PCA is defined as an orthogonal linear transformation on a real inner product space that transforms the data to a new coordinate system such that the greatest variance by some scalar projection of the data comes to lie on the first coordinate (called the first principal component), the second greatest variance on the second coordinate, and so on. Consider an n × p {\displaystyle n\times p} data matrix , X , with column-wise zero empirical mean (the sample mean of each column has been shifted to zero), where each of

10812-517: The aim of finding long term oscillations of natural variability in global climate. The resulting reconstruction went back to 1400, and was published in November as Mann, Park & Bradley 1995 . They were able to detect that the multiple proxies were varying in a coherent oscillatory way, indicating both the multidecadal pattern in the North Atlantic and a longer term oscillation of roughly 250 years in

10971-467: The aspect that got the most attention. Their original draft ended in 1980 as most reconstructions only went that far, but an anonymous peer reviewer of the paper suggested that the curve of instrumental temperature records should be shown up to the present to include the considerable warming that had taken place between 1980 and 1998. The Mann, Bradley & Hughes 1998 (MBH98) multiproxy study on "Global-scale temperature patterns and climate forcing over

11130-478: The back of an envelope", a "rather dodgy bit of hand-waving". In Bradley 1991 , a working group of climatologists including Raymond S. Bradley , Malcolm K. Hughes , Jean Jouzel , Wibjörn Karlén , Jonathan Overpeck and Tom Wigley proposed a project to improve understanding of natural climatic variations over the last two thousand years so that their effect could be allowed for when evaluating human contributions to climate change. Climate proxy temperature data

11289-456: The basis of their compilation of 20 time-series. These records were largely instrumental but also included some proxy records including two tree-ring series. Their method used nested multiple regression to allow for records covering different periods, and produced measures of uncertainty. The reconstruction showed a cool period extending beyond the Maunder Minimum , and warmer temperatures in

11448-468: The borehole reconstruction published by Pollack, Huang and Shen gave independent support to the conclusion that 20th century warmth was exceptional for the past 500 years. Jones, Keith Briffa , Tim P. Barnett and Simon Tett had independently produced a "Composite Plus Scale" (CPS) reconstruction extending back for a thousand years, comparing tree ring, coral layer, and glacial proxy records, but not specifically estimating uncertainties. Jones et al. 1998

11607-573: The broad consensus shown in the original 1998 hockey-stick graph, with variations in how flat the pre-20th century "shaft" appears. The 2007 IPCC Fourth Assessment Report cited 14 reconstructions, 10 of which covered 1,000 years or longer, to support its strengthened conclusion that it was likely that Northern Hemisphere temperatures during the 20th century were the highest in at least the past 1,300 years. Further reconstructions, including Mann et al. 2008 and PAGES 2k Consortium 2013 , have supported these general conclusions. Paleoclimatology influenced

11766-703: The case for a thousand or more years". Tim Barnett of the Scripps Institution of Oceanography was working towards the next IPCC assessment with Phil Jones , and in 1996 told journalist Fred Pearce "What we hope is that the current patterns of temperature change prove distinctive, quite different from the patterns of natural variability in the past". A divergence problem affecting some tree ring proxies after 1960 had been identified in Alaska by Taubes 1995 and Jacoby & d'Arrigo 1995 . Tree ring specialist Keith Briffa 's February 1998 study showed that this problem

11925-440: The case of speed (i.e., distance per unit of time): For example, the harmonic mean of the five values: 4, 36, 45, 50, 75 is If we have five pumps that can empty a tank of a certain size in respectively 4, 36, 45, 50, and 75 minutes, then the harmonic mean of 15 {\displaystyle 15} tells us that these five different pumps working together will pump at the same rate as much as five pumps that can each empty

12084-461: The claim that all of the warming took place between 1920 and 1935, before increased human greenhouse gas emissions. The George C. Marshall Institute alleged that MBH98 was deceptive in only going back to 1400, and so not covering the Medieval Warm Period which predated industrial greenhouse gas emissions. The same criticisms were made by Willie Soon and Sallie Baliunas . In October 1998

12243-465: The climate of the region or hemisphere over time. This method was implemented in various ways, including different selection processes for the proxy records, and averaging could be unweighted, or could be weighted in relation to an assessment of reliability or of area represented. There were also different ways of finding the scaling coefficient used to scale the proxy records to the instrumental temperature record. John A. Eddy had earlier tried to relate

12402-428: The climate system varied and responded to radiative forcing , so hemispheric averages were a secondary product. The CFR method could also be used to reconstruct Northern Hemisphere mean temperatures, and the results closely resembled the earlier CPS reconstructions including Bradley & Jones 1993 . Mann describes this as the least scientifically interesting thing they could do with the rich spatial patterns, but also

12561-636: The conclusion that peak Medieval warmth only occurred during two or three short periods of 20 to 30 years, with temperatures around 1950s levels, refuting claims that 20th century warming was not unusual. An analysis by Crowley published in July 2000 compared simulations from an energy balance climate model with reconstructed mean annual temperatures from MBH99 and Crowley & Lowery (2000). While earlier reconstructed temperature variations were consistent with volcanic and solar irradiation changes plus residual variability, very large 20th-century warming closely agreed with

12720-476: The controversy, the publisher of Climate Research upgraded Hans von Storch from editor to editor in chief, but von Storch decided that the Soon and Baliunas paper was seriously flawed and should not have been published as it was. He proposed a new editorial system, and though the publisher of Climate Research agreed that the paper should not have been published uncorrected, he rejected von Storch's proposals to improve

12879-509: The corresponding singular value. This form is also the polar decomposition of T . Efficient algorithms exist to calculate the SVD of X without having to form the matrix X X , so computing the SVD is now the standard way to calculate a principal components analysis from a data matrix, unless only a handful of components are required. As with the eigen-decomposition, a truncated n × L score matrix T L can be obtained by considering only

13038-410: The data matrix. PCA is the simplest of the true eigenvector-based multivariate analyses and is closely related to factor analysis . Factor analysis typically incorporates more domain-specific assumptions about the underlying structure and solves eigenvectors of a slightly different matrix. PCA is also related to canonical correlation analysis (CCA) . CCA defines coordinate systems that optimally describe

13197-877: The data set successively inherit the maximum possible variance from X , with each coefficient vector w constrained to be a unit vector (where l {\displaystyle l} is usually selected to be strictly less than p {\displaystyle p} to reduce dimensionality). The above may equivalently be written in matrix form as where T i k = t k ( i ) {\displaystyle {\mathbf {T} }_{ik}={t_{k}}_{(i)}} , X i j = x j ( i ) {\displaystyle {\mathbf {X} }_{ij}={x_{j}}_{(i)}} , and W j k = w j ( k ) {\displaystyle {\mathbf {W} }_{jk}={w_{j}}_{(k)}} . In order to maximize variance,

13356-524: The data used in MBH98 paper. In 2004 Hans von Storch published criticism of the statistical techniques as tending to underplay variations in earlier parts of the graph, though this was disputed and he later accepted that the effect was very small. In 2005 McIntyre and McKitrick published criticisms of the principal components analysis methodology as used in MBH98 and MBH99. Their analysis was subsequently disputed by published papers including Huybers 2005 and Wahl & Ammann 2007 which pointed to errors in

13515-432: The data. This choice of basis will transform the covariance matrix into a diagonalized form, in which the diagonal elements represent the variance of each axis. The proportion of the variance that each eigenvector represents can be calculated by dividing the eigenvalue corresponding to that eigenvector by the sum of all eigenvalues. Biplots and scree plots (degree of explained variance ) are used to interpret findings of

13674-402: The dataset X . The sample covariance Q between two of the different principal components over the dataset is given by: where the eigenvalue property of w ( k ) has been used to move from line 2 to line 3. However eigenvectors w ( j ) and w ( k ) corresponding to eigenvalues of a symmetric matrix are orthogonal (if the eigenvalues are different), or can be orthogonalised (if

13833-433: The dataset is not too large, the significance of the principal components can be tested using parametric bootstrap , as an aid in determining how many principal components to retain. The principal components transformation can also be associated with another matrix factorization, the singular value decomposition (SVD) of X , Here Σ is an n -by- p rectangular diagonal matrix of positive numbers σ ( k ) , called

13992-480: The debate, and Broecker's criticism that MBH99 did not show a clear MWP. They concluded that the MWP was likely to have been widespread in the extratropical northern hemisphere, and seemed to have approached late 20th century temperatures at times. In an interview, Mann said the study did not contradict MBH as it dealt only with extratropical land areas, and stopped before the late 20th century. He reported that Edward R. Cook ,

14151-523: The distinction of being the optimal orthogonal transformation for keeping the subspace that has largest "variance" (as defined above). This advantage, however, comes at the price of greater computational requirements if compared, for example, and when applicable, to the discrete cosine transform , and in particular to the DCT-II which is simply known as the "DCT". Nonlinear dimensionality reduction techniques tend to be more computationally demanding than PCA. PCA

14310-561: The dominant climate forcing during the 20th century. In a review in the same issue, Gabriele C. Hegerl described their method as "quite original and promising", which could help to verify model estimates of natural climate fluctuations and was "an important step towards reconstructing space–time records of historical temperature patterns". Release of the paper on 22 April 1998 was given exceptional media coverage, including questioning as to whether it proved that human influences were responsible for climate change . Mann would only agree that it

14469-406: The earlier period. The MBH99 conclusion that the 1990s were likely to have been the warmest decade, and 1998 the warmest year, of the past millennium in the Northern Hemisphere, with "likely" defined as "66-90% chance", was supported by reconstructions by Crowley & Lowery 2000 and by Jones et al. 1998 using different data and methods. The Pollack, Huang & Shen 1998 reconstruction covering

14628-571: The editorial process, and von Storch with three other board members resigned. Senator James M. Inhofe stated his belief that "manmade global warming is the greatest hoax ever perpetrated on the American people", and a hearing of the United States Senate Committee on Environment and Public Works which he convened on 29 July 2003 heard the news of the resignations. Mean The arithmetic mean , also known as "arithmetic average",

14787-687: The effect its removal had on the result. He found that certain proxies were critical to the reliability of the reconstruction, particularly one tree ring dataset collected by Gordon Jacoby and Rosanne D'Arrigo in a part of North America Bradley's earlier research had identified as a key region. This dataset only extended back to 1400, and though another proxy dataset from the same region (in the International Tree-Ring Data Bank ) went further back and should have given reliable proxies for earlier periods, validation tests only supported their reconstruction after 1400. To find out why, Mann compared

14946-410: The first i − 1 {\displaystyle i-1} principal components that maximizes the variance of the projected data. For either objective, it can be shown that the principal components are eigenvectors of the data's covariance matrix . Thus, the principal components are often computed by eigendecomposition of the data covariance matrix or singular value decomposition of

15105-463: The first eigenvector -based climate field reconstruction (CFR). This showed global patterns of annual surface temperature, and included a graph of average hemispheric temperatures back to 1400 with shading emphasising that uncertainties (to two standard error limits) were much greater in earlier centuries. Jones et al. 1998 independently produced a CPS reconstruction extending back for a thousand years, and Mann, Bradley & Hughes 1999 (MBH99) used

15264-411: The first k  − 1 principal components from X : and then finding the weight vector which extracts the maximum variance from this new data matrix It turns out that this gives the remaining eigenvectors of X X , with the maximum values for the quantity in brackets given by their corresponding eigenvalues. Thus the weight vectors are eigenvectors of X X . The k -th principal component of

15423-500: The first L largest singular values and their singular vectors: The truncation of a matrix M or T using a truncated singular value decomposition in this way produces a truncated matrix that is the nearest possible matrix of rank L to the original matrix, in the sense of the difference between the two having the smallest possible Frobenius norm , a result known as the Eckart–Young theorem [1936]. The singular values (in Σ ) are

15582-456: The first few principal components compared to the same noise variance, the proportionate effect of the noise is less—the first few components achieve a higher signal-to-noise ratio . PCA thus can have the effect of concentrating much of the signal into the first few principal components, which can usefully be captured by dimensionality reduction; while the later principal components may be dominated by noise, and so disposed of without great loss. If

15741-430: The first principal component of a set of p {\displaystyle p} variables is the derived variable formed as a linear combination of the original variables that explains the most variance. The second principal component explains the most variance in what is left once the effect of the first component is removed, and we may proceed through p {\displaystyle p} iterations until all

15900-415: The first weight vector w (1) thus has to satisfy Equivalently, writing this in matrix form gives Since w (1) has been defined to be a unit vector, it equivalently also satisfies The quantity to be maximised can be recognised as a Rayleigh quotient . A standard result for a positive semidefinite matrix such as X X is that the quotient's maximum possible value is the largest eigenvalue of

16059-608: The founding director of the Climatic Research Unit (CRU) in the University of East Anglia (UEA), which aimed to improve knowledge of climate history in both the recent and far distant past, monitor current changes in global climate, identify processes causing changes at different timescales, and review the possibility of advising about future trends in climate. During the cold years of the 1960s, Lamb had anticipated that natural cycles were likely to lead over thousands of years to

16218-422: The globe, including a rich resource of tree ring networks for some areas and sparser proxies such as lake sediments, ice cores and corals, as well as some historical records. Their global reconstruction was a major breakthrough in evaluation of past climate dynamics, and the first eigenvector -based climate field reconstruction (CFR) incorporating multiple climate proxy data sets of different types and lengths into

16377-452: The important step of validation calculations , which showed that the reconstructions were statistically meaningful, or skillful . A balance was required over the whole globe, but most of the proxy data came from tree rings in the Northern mid latitudes , largely in dense proxy networks. Since using all of the large numbers of tree ring records in would have overwhelmed the sparse proxies from

16536-436: The instrumental record from the proxy evidence and emphasising the increasing range of possible error in earlier times, which MBH said would "preclude, as yet, any definitive conclusions" about climate before 1400. The reconstruction found significant variability around a long-term cooling trend of –0.02 °C per century, as expected from orbital forcing , interrupted in the 20th century by rapid warming which stood out from

16695-484: The instrumental record. The least squares simultaneous solution of these multiple regressions used covariance between the proxy records. The results were then used to reconstruct large-scale patterns over time in the spatial field of interest (defined as the empirical orthogonal functions , or EOFs) using both local relationships of the proxies to climate and distant climate teleconnections . Temperature records for almost 50 years prior to 1902 were analysed using PCA for

16854-443: The integral converges. But the mean may be finite even if the function itself tends to infinity at some points. Angles , times of day, and other cyclical quantities require modular arithmetic to add and otherwise combine numbers. In all these situations, there will not be a unique mean. For example, the times an hour before and after midnight are equidistant to both midnight and noon. It is also possible that no mean exists. Consider

17013-521: The last 1,000 years. In the Soon and Baliunas controversy , two scientists cited in the papers said that their work was misrepresented, and the Climate Research paper was criticised by many other scientists, including several of the journal's editors. On 8 July Eos featured a detailed rebuttal of both papers by 13 scientists including Mann and Jones, presenting strong evidence that Soon and Baliunas had used improper statistical methods. Responding to

17172-411: The lowest and the highest quarter of values. assuming the values have been ordered, so is simply a specific example of a weighted mean for a specific set of weights. In some circumstances, mathematicians may calculate a mean of an infinite (or even an uncountable ) set of values. This can happen when calculating the mean value y avg {\displaystyle y_{\text{avg}}} of

17331-409: The matrix, which occurs when w is the corresponding eigenvector . With w (1) found, the first principal component of a data vector x ( i ) can then be given as a score t 1( i ) = x ( i ) ⋅ w (1) in the transformed co-ordinates, or as the corresponding vector in the original variables, { x ( i ) ⋅ w (1) } w (1) . The k -th component can be found by subtracting

17490-422: The mean and size of sample i {\displaystyle i} respectively. In other applications, they represent a measure for the reliability of the influence upon the mean by the respective values. Sometimes, a set of numbers might contain outliers (i.e., data values which are much lower or much higher than the others). Often, outliers are erroneous data caused by artifacts . In this case, one can use

17649-460: The mean for a moderately skewed distribution. It is used in hydrocarbon exploration and is defined as: where P 10 {\textstyle P_{10}} , P 50 {\textstyle P_{50}} and P 90 {\textstyle P_{90}} are the 10th, 50th and 90th percentiles of the distribution, respectively. Principal components analysis Principal component analysis ( PCA )

17808-424: The millennium, with 1998 the warmest year so far." Bradley was quoted as saying "Temperatures in the latter half of the 20th century were unprecedented", while Mann said "As you go back farther in time, the data becomes sketchier. One can't quite pin things down as well, but, our results do reveal that significant changes have occurred, and temperatures in the latter 20th century have been exceptionally warm compared to

17967-399: The new multivariate method of relating these series to the instrumental data is as good as the paper claims, it should be statistically reliable." He discussed some of the difficulties, and emphasised that "Each paleoclimatic discipline has to come to terms with its own limitations and must unreservedly admit to problems, warts and all." The study was disputed by contrarian Pat Michaels with

18126-499: The northern hemisphere average, and compared the instrumental record with the proxy reconstruction over the same period. It concluded that the "Little Ice Age" period was complex, with evidence suggesting the influence of volcanic eruptions. It showed that temperatures since the 1920s were higher than earlier in the 500-year period, an indication of other factors which could most probably be attributed to human caused changes increasing levels of greenhouse gases . This paper introduced

18285-415: The original observed values for each of the variables. Then, we compute the covariance matrix of the data and calculate the eigenvalues and corresponding eigenvectors of this covariance matrix. Then we must normalize each of the orthogonal eigenvectors to turn them into unit vectors. Once this is done, each of the mutually-orthogonal unit eigenvectors can be interpreted as an axis of the ellipsoid fitted to

18444-425: The original variables can be written The empirical covariance matrix between the principal components becomes where Λ is the diagonal matrix of eigenvalues λ ( k ) of X X . λ ( k ) is equal to the sum of the squares over the dataset associated with each component k , that is, λ ( k ) = Σ i t k ( i ) = Σ i ( x ( i ) ⋅ w ( k ) ) . The transformation T = X W maps

18603-517: The parameter m , the following types of means are obtained: This can be generalized further as the generalized f -mean and again a suitable choice of an invertible f will give The weighted arithmetic mean (or weighted average) is used if one wants to combine average values from different sized samples of the same population: Where x i ¯ {\displaystyle {\bar {x_{i}}}} and w i {\displaystyle w_{i}} are

18762-428: The past 500 years gave independent support for this conclusion, which was compared against the independent (extra-tropical, warm-season) tree-ring density NH temperature reconstruction of Briffa 2000 . Its Figure 2.21 showed smoothed curves from the MBH99, Jones et al. and Briffa reconstructions, together with modern thermometer data as a red line and the grey shaded 95% confidence range from MBH99. Above it, figure 2.20

18921-428: The past five centuries supported the conclusion that 20th century warming was exceptional. In a perspective commenting on MBH99, Wallace Smith Broecker argued that the Medieval Warm Period (MWP) was global. He attributed recent warming to a roughly 1500-year cycle which he suggested related to episodic changes in the Atlantic's conveyor circulation . A March 2002 tree ring reconstruction by Jan Esper et al. noted

19080-407: The past millennium: inferences, uncertainties, and limitations to emphasise the increasing uncertainty involved in reconstructions of the period before 1400 when fewer proxies were available. A University of Massachusetts Amherst news release dated 3 March 1999 announced publication in the 15 March issue of Geophysical Research Letters , "strongly suggesting that the 1990s were the warmest decade of

19239-532: The past six centuries" was submitted to the journal Nature on 9 May 1997, accepted on 27 February 1998 and published on 23 April 1998. The paper announced a new statistical approach to find patterns of climate change in both time and global distribution, building on previous multiproxy reconstructions. The authors concluded that "Northern Hemisphere mean annual temperatures for three of the past eight years are warmer than any other year since (at least) AD1400", and estimated empirically that greenhouse gases had become

19398-472: The point that bristlecone pines from the Western U.S. could have been affected by pollution such as rising CO 2 levels as well as temperature. The temperature curve was supported by other studies, but most of these shared the limited well dated proxy evidence then available, and so few were truly independent. The uncertainties in earlier times rose as high as those in the reconstruction at 1980, but did not reach

19557-515: The points to the line . These directions (i.e., principal components) constitute an orthonormal basis in which different individual dimensions of the data are linearly uncorrelated . Many studies use the first two principal components in order to plot the data in two dimensions and to visually identify clusters of closely related data points. Principal component analysis has applications in many fields such as population genetics , microbiome studies, and atmospheric science . When performing PCA,

19716-511: The power mean or Hölder mean, is an abstraction of the quadratic , arithmetic, geometric, and harmonic means. It is defined for a set of n positive numbers x i by x ¯ ( m ) = ( 1 n ∑ i = 1 n x i m ) 1 m {\displaystyle {\bar {x}}(m)=\left({\frac {1}{n}}\sum _{i=1}^{n}x_{i}^{m}\right)^{\frac {1}{m}}} By choosing different values for

19875-415: The preceding 900 years. Though substantial uncertainties exist in the estimates, these are nonetheless startling revelations." While the reconstruction supported theories of a relatively warm medieval period, Hughes said "even the warmer intervals in the reconstruction pale in comparison with mid-to-late 20th-century temperatures." The New York Times report had a colored version of the graph, distinguishing

20034-422: The predicted effects of greenhouse gas emissions. Reviewing twenty years of progress in palaeoclimatology, Jones noted the reconstructions by Jones et al. (1998), MBH99, Briffa (2000) and Crowley & Lowery (2000) showing good agreement using different methods, but cautioned that use of many of the same proxy series meant that they were not independent, and more work was needed. The Working Group 1 (WG1) part of

20193-426: The rarity of sunspots during the Maunder Minimum to Lamb's estimates of past climate, but had insufficient information to produce a quantitative assessment. The problem was reexamined by Bradley in collaboration with solar physicists Judith Lean and Juerg Beer , using the findings of Bradley & Jones 1993 . The Lean, Beer & Bradley 1995 paper confirmed that the drop in solar output appeared to have caused

20352-457: The reconstruction procedure was used with these pseudoproxies, the result was then compared with the original record or simulation to see how closely it had been reconstructed. The paper discussed the issue that regression methods of reconstruction tended to underestimate the amplitude of variation. While the IPCC Third Assessment Report (TAR) drew on five reconstructions to support its conclusion that recent Northern Hemisphere temperatures were

20511-503: The reconstruction useful for investigating natural variability and long-term oscillations as well as for comparisons with patterns produced by climate models. The CFR method made more use of climate information embedded in remote proxies, but was more dependent than CPS on assumptions that relationships between proxy indicators and large-scale climate patterns remained stable over time. Related rigorous statistical methods had been developed for tree ring data, with Harold C. Fritts publishing

20670-431: The reconstructions have been taken up by fossil fuel industry funded lobbying groups attempting to cast doubt on climate science. Paleoclimatology dates back to the 19th century, and the concept of examining varves in lake beds and tree rings to track local climatic changes was suggested in the 1930s. In the 1960s, Hubert Lamb generalised from historical documents and temperature records of central England to propose

20829-556: The region to produce a corrected version of this dataset. Their reconstruction using this corrected dataset passed the validation tests for the extended period, but they were cautious about the increased uncertainties. The Mann, Bradley and Hughes reconstruction covering 1,000 years (MBH99) was submitted in October 1998 to Geophysical Research Letters which published it in March 1999 with the cautious title Northern Hemisphere temperatures during

20988-430: The right singular vectors W of X are equivalent to the eigenvectors of X X , while the singular values σ ( k ) of X {\displaystyle \mathbf {X} } are equal to the square-root of the eigenvalues λ ( k ) of X X . Using the singular value decomposition the score matrix T can be written so each column of T is given by one of the left singular vectors of X multiplied by

21147-430: The same as that variable, with a small contribution from the other variable, whereas the second component will be almost aligned with the second original variable. This means that whenever the different variables have different units (like temperature and mass), PCA is a somewhat arbitrary method of analysis. (Different results would be obtained if one used Fahrenheit rather than Celsius for example.) Pearson's original paper

21306-407: The set of points around their mean in order to align with the principal components. This moves as much of the variance as possible (using an orthogonal transformation) into the first few dimensions. The values in the remaining dimensions, therefore, tend to be small and may be dropped with minimal loss of information (see below ). PCA is often used in this manner for dimensionality reduction . PCA has

21465-466: The singular values of X ; U is an n -by- n matrix, the columns of which are orthogonal unit vectors of length n called the left singular vectors of X ; and W is a p -by- p matrix whose columns are orthogonal unit vectors of length p and called the right singular vectors of X . In terms of this factorization, the matrix X X can be written where Σ ^ {\displaystyle \mathbf {\hat {\Sigma }} }

21624-415: The square roots of the eigenvalues of the matrix X X . Each eigenvalue is proportional to the portion of the "variance" (more correctly of the sum of the squared distances of the points from their multidimensional mean) that is associated with each eigenvector. The sum of all the eigenvalues is equal to the sum of the squared distances of the points from their multidimensional mean. PCA essentially rotates

21783-467: The sum is taken over all possible values of the random variable and P ( x ) {\displaystyle P(x)} is the probability mass function . For a continuous distribution , the mean is ∫ − ∞ ∞ x f ( x ) d x {\displaystyle \textstyle \int _{-\infty }^{\infty }xf(x)\,dx} , where f ( x ) {\displaystyle f(x)}

21942-544: The surrounding region. Their study did not calibrate these proxy patterns against a quantitative temperature scale, and a new statistical approach was needed to find how they related to surface temperatures in order to reconstruct past temperature patterns. For his postdoctoral research Mann joined Bradley and tree ring specialist Malcolm K. Hughes to develop a new statistical approach to reconstruct underlying spatial patterns of temperature variation combining diverse datasets of proxy information covering different periods across

22101-419: The tank in 15 {\displaystyle 15} minutes. AM, GM, and HM satisfy these inequalities: Equality holds if all the elements of the given sample are equal. In descriptive statistics , the mean may be confused with the median , mode or mid-range , as any of these may incorrectly be called an "average" (more formally, a measure of central tendency ). The mean of a set of observations

22260-497: The temperatures of later thermometer data. They concluded that although the 20th century was almost certainly the warmest of the millennium, the amount of anthropogenic warming remains uncertain." With work progressing on the next IPCC report, Chris Folland told researchers on 22 September 1999 that a figure showing temperature changes over the millennium "is a clear favourite for the policy makers' summary". Two graphs competed: Jones et al. (1998) and MBH99. In November, Jones produced

22419-523: The two datasets and found that they tracked each other closely from 1400 to 1800, then diverged until around 1900 when they again tracked each other. He found a likely reason in the CO 2 " fertilisation effect " affecting tree rings as identified by Graybill and Idso, with the effect ending once CO 2 levels had increased to the point where warmth again became the key factor controlling tree growth at high altitude. Mann used comparisons with other tree ring data from

22578-428: The variables in a dataset are noisy. If each column of the dataset contains independent identically distributed Gaussian noise, then the columns of T will also contain similarly identically distributed Gaussian noise (such a distribution is invariant under the effects of the matrix W , which can be thought of as a high-dimensional rotation of the co-ordinate axes). However, with more of the total variance concentrated in

22737-422: The variance is explained. PCA is most commonly used when many of the variables are highly correlated with each other and it is desirable to reduce their number to an independent set. The first principal component can equivalently be defined as a direction that maximizes the variance of the projected data. The i {\displaystyle i} -th principal component can be taken as a direction orthogonal to

22896-421: The vectors happen to share an equal repeated value). The product in the final line is therefore zero; there is no sample covariance between different principal components over the dataset. Another way to characterise the principal components transformation is therefore as the transformation to coordinates which diagonalise the empirical sample covariance matrix. In matrix form, the empirical covariance matrix for

23055-405: The warmest in the past 1,000 years, it gave particular prominence to an IPCC illustration based on the MBH99 paper. The hockey stick graph was subsequently seen by mass media and the public as central to the IPCC case for global warming, which had actually been based on other unrelated evidence. From an expert viewpoint the graph was, like all newly published science, preliminary and uncertain, but it

23214-433: The whole period, with the 1990s "the warmest decade, and 1998 the warmest year, at moderately high levels of confidence." This was illustrated by the time series line graph Figure 2(a) which showed their reconstruction from AD 1000 to 1980 as a thin line, wavering around a thicker dark 40-year smoothed line. This curve followed a downward trend (shown as a thin dot-dashed line) from a Medieval Warm Period (about as warm as

23373-421: Was criticised as misleading . Briffa's paper as published in the January 2000 issue of Quaternary Science Reviews showed the unusual warmth of the last century, but cautioned that the impact of human activities on tree growth made it subtly difficult to isolate a clear climate message. In February 2000 Thomas J. Crowley and Thomas S. Lowery 's reconstruction incorporated data not used previously. It reached

23532-423: Was "highly suggestive" of that inference. He said that "Our conclusion was that the warming of the past few decades appears to be closely tied to emission of greenhouse gases by humans and not any of the natural factors". Most proxy data are inherently imprecise, and Mann said "We do have error bars. They are somewhat sizable as one gets farther back in time, and there is reasonable uncertainty in any given year. There

23691-409: Was "not yet sufficiently easy to assess nor sufficiently integrated with indications from other data to be used in this report." A "schematic diagram" of global temperature variations over the last thousand years has been traced to a graph based loosely on Lamb's 1965 paper, nominally representing central England, modified by Lamb in 1982. Mike Hulme describes this schematic diagram as "Lamb's sketch on

23850-531: Was adapted from MBH99. Figure 5 in WG1 Technical Summary B (as shown to the right) repeated this figure without the linear trend line declining from AD 1000 to 1850. This iconic graph adapted from MBH99 was featured prominently in the WG1 Summary for Policymakers under a graph of the instrumental temperature record for the past 140 years. The text stated that it was "likely that, in the Northern Hemisphere,

24009-461: Was anomalous over the 300-year period, and went as far as speculating that these results supported the hypothesis that recent warming had human causes. Publicity over the concerns of scientists about the implications of global warming led to increasing public and political interest, and the Reagan administration , concerned in part about the political impact of scientific findings, successfully lobbied for

24168-466: Was entitled "On Lines and Planes of Closest Fit to Systems of Points in Space" – "in space" implies physical Euclidean space where such concerns do not arise. One way of making the PCA less arbitrary is to use variables scaled so as to have unit variance, by standardizing the data and hence use the autocorrelation matrix instead of the autocovariance matrix as a basis for PCA. However, this compresses (or expands)

24327-482: Was exceptional. Bradley & Jones 1993 introduced the "Composite Plus Scaling" (CPS) method which, as of 2009, was still being used by most large-scale reconstructions. Their study was featured in the IPCC Second Assessment Report of 1995. In 1998 Michael E. Mann , Raymond S. Bradley and Malcolm K. Hughes developed new statistical techniques to produce Mann, Bradley & Hughes 1998 (MBH98),

24486-445: Was more widespread at high northern latitudes, and warned that it had to be taken into account to avoid overestimating past temperatures. Variations on the "Composite Plus Scale" (CPS) method continued to be used to produce hemispheric or global mean temperature reconstructions. From 1998 this was complemented by Climate Field Reconstruction (CFR) methods which could show how climate patterns had developed over large spatial areas, making

24645-467: Was needed at seasonal or annual resolution covering a wide geographical area to provide a framework for testing the part climate forcings had played in past variations, look for cycles in climate, and find if debated climatic events such as the Little Ice Age and Medieval Warm Period were global. Reconstructions were to be made of key climate systems, starting with three climatically sensitive regions:

24804-512: Was not unusual. In March they published an extended paper in Energy & Environment , with additional authors. The Bush administration 's Council on Environmental Quality chief of staff Philip Cooney inserted references to the papers in the draft first Environmental Protection Agency Report on the Environment , and removed all references to reconstructions showing world temperatures rising over

24963-483: Was popularized by the climatologist Jerry Mahlman , to describe the pattern shown by the Mann, Bradley & Hughes 1999 (MBH99) reconstruction, envisaging a graph that is relatively flat with a downward trend to 1900 as forming an ice hockey stick 's "shaft" followed by a sharp, steady increase corresponding to the "blade" portion. The reconstructions have featured in Intergovernmental Panel on Climate Change (IPCC) reports as evidence of global warming . Arguments over

25122-545: Was submitted to The Holocene on 16 October 1997; their revised manuscript was accepted on 3 February and published in May 1998. As Bradley recalls, Mann's initial view was that there was too little information and too much uncertainty to go back so far, but Bradley said "Why don't we try to use the same approach we used in Nature, and see if we could push it back a bit further?" Within a few weeks, Mann responded that to his surprise, "There

25281-489: Was widely used to publicise the issue of global warming, and it was targeted by those opposing ratification of the Kyoto Protocol on global warming. A literature review by Willie Soon and Sallie Baliunas , published in the relatively obscure journal Climate Research on 31 January 2003, used data from previous papers to argue that the Medieval Warm Period had been warmer than the 20th century, and that recent warming

#346653