Although there is broad consensus that there will be a significant rise in average global temperature, there is great uncertainty over the extent of the change, and the implications for different regions. Greater accuracy is urgently needed to provide a sound basis for major policy decisions and to ensure that politicians and the public remain convinced that significant changes in consumption patterns and energy production are essential to stave off serious consequences in the coming decades and centuries.
The climate modelling community has become increasingly aware that some of the statistical tools that could improve their modelling of climate change may already have been developed for econometric problems, which have some of the same features. The European Science Foundation (ESF) brought these two communities together for the first time in a recent workshop, sowing the seeds for future collaboration.
“We achieved our goal of bringing together people from two very distant but equally valuable fields,” said the workshop’s co-convenor Peter Thejll. “It was designed as a one-way session whereby econometricians were supposed to convey knowledge of econometric methods to the climate researchers.”
This has already proved highly valuable because economic and climate models require similar kinds of statistical analysis, both for example involving serial correlation where the aim is to predict the future value of a variable based upon a starting value at an earlier point in time. In economics such a variable might be the price of a commodity, while in climatology it might be temperature or atmospheric pressure. In both cases the variables change randomly during successive time intervals subject to varying constraints within a closely defined zone, and therefore can be analysed using similar “random walk” techniques.
“To solve important climate problems related to climate change and change attribution with statistics, these methods have to be used and understood by climate researchers,” said Thejll. “We brought together people who understand these problems and had a great, and informative, time.”
Thejll is confident the new found cooperation with the econometric community will improve climate modelling and forecasting, but first there is a need to digest some of the new tools and ideas. The aim is to introduce greater statistical sophistication into climate analysis, partly by understanding better the correlation between different aspects of change, for example how one region impacts another. “We first need to see the spread of econometric methods so that we no longer read climate research papers that ignore important statistical problems,” said Thejll wryly.
This will lead to an important first step towards better climate change predictions models – understanding the limitations of existing models. “One improvement that can follow from the use of econometric methods in climate research is a better understanding of the level of ignorance we have,” said Thejll.
One problem at present is that uncertainties are commonly underestimated, and this makes it very difficult to predict with much confidence even the broad climatic consequences or rising atmospheric carbon dioxide levels. But Thejll hopes and expects that by incorporating the key tools of econometric modelling, climate prediction will become much more accurate and valuable.
The ESF workshop, Econometric Time Series Analysis applied to climate research, was held in Frascati, Italy, in September 2007.Contact:
Dispersal of Fish Eggs by Water Birds – Just a Myth?
19.02.2018 | Universität Basel
Removing fossil fuel subsidies will not reduce CO2 emissions as much as hoped
08.02.2018 | International Institute for Applied Systems Analysis (IIASA)
Study published in the journal ACS Applied Materials & Interfaces is the outcome of an international effort that included teams from Dresden and Berlin in Germany, and the US.
Scientists at the Helmholtz-Zentrum Dresden-Rossendorf (HZDR) together with colleagues from the Helmholtz-Zentrum Berlin (HZB) and the University of Virginia...
Novel highly efficient and brilliant gamma-ray source: Based on model calculations, physicists of the Max PIanck Institute for Nuclear Physics in Heidelberg propose a novel method for an efficient high-brilliance gamma-ray source. A giant collimated gamma-ray pulse is generated from the interaction of a dense ultra-relativistic electron beam with a thin solid conductor. Energetic gamma-rays are copiously produced as the electron beam splits into filaments while propagating across the conductor. The resulting gamma-ray energy and flux enable novel experiments in nuclear and fundamental physics.
The typical wavelength of light interacting with an object of the microcosm scales with the size of this object. For atoms, this ranges from visible light to...
Stable joint cartilage can be produced from adult stem cells originating from bone marrow. This is made possible by inducing specific molecular processes occurring during embryonic cartilage formation, as researchers from the University and University Hospital of Basel report in the scientific journal PNAS.
Certain mesenchymal stem/stromal cells from the bone marrow of adults are considered extremely promising for skeletal tissue regeneration. These adult stem...
In the fight against cancer, scientists are developing new drugs to hit tumor cells at so far unused weak points. Such a “sore spot” is the protein complex...
In an article that appears in the journal “Review of Modern Physics”, researchers at the Laboratory for Attosecond Physics (LAP) assess the current state of the field of ultrafast physics and consider its implications for future technologies.
Physicists can now control light in both time and space with hitherto unimagined precision. This is particularly true for the ability to generate ultrashort...
13.04.2018 | Event News
12.04.2018 | Event News
09.04.2018 | Event News
19.04.2018 | Materials Sciences
19.04.2018 | Physics and Astronomy
19.04.2018 | Physics and Astronomy