Americans face 70 percent increase in unhealthy ozone levels by 2050
Ozone pollution across the continental United States will become far more difficult to keep in check as temperatures rise, according to new research results.
The study shows that Americans face the risk of a 70 percent increase in unhealthy summertime ozone levels by 2050.
The results appear online this week in a paper in the Journal of Geophysical Research-Atmospheres, published by the American Geophysical Union.
The work was funded by the National Science Foundation (NSF) and the U.S. Department of Energy.
Warmer temperatures and other changes in the atmosphere related to a changing climate, including higher atmospheric levels of methane, spur chemical reactions that increase overall levels of ozone.
Unlike ozone in the stratosphere, which benefits life on Earth by blocking ultraviolet radiation from the sun, ground-level ozone can trigger a number of health problems.
These range from coughing and throat irritation to more serious aggravation of asthma, bronchitis and emphysema.
Even short periods of unhealthy ozone levels can cause local death rates to rise. Ozone pollution also damages crops and other plants.
Unless emissions of specific pollutants associated with the formation of ozone are sharply cut, most of the continental United States will experience more summer days with unhealthy air by 2050, the research shows.
Heavily polluted locations in parts of the East, Midwest and West Coast, in which ozone already frequently exceeds recommended levels, could face unhealthy summer air in most years.
"It doesn't matter where you are in the United States, climate change has the potential to make your air worse," said National Center for Atmospheric Research (NCAR) scientist Gabriele Pfister, lead scientist on the study.
In addition to NCAR, the paper co-authors are from the Pacific Northwest National Laboratory; University of Colorado, Boulder; and North-West University in South Africa.
"A warming planet doesn't just mean rising temperatures, it also means risking more summertime pollution and the health effects that come with it," said Pfister.
However, the research also showed that a sharp reduction in the emissions of certain pollutants would lead to dramatically decreased levels of ozone even as temperatures warm.
The research is one of the first of its type to be conducted with new, highly advanced geoscience supercomputing capabilities.
"Understanding future changes in surface ozone over the summer has tremendous implications for air quality and human health," said Anjuli Bamzai, a program director in NSF's Division of Atmospheric and Geospace Sciences, which funded the research through NSF's Decadal and Regional Climate Prediction using Earth System Models (EaSM) Program.
"Through a series of 'what if' simulations," said Bamzai, "atmospheric chemists, climate modelers, regional modelers and developers of emissions scenarios demonstrate that a balance of emission controls can counteract the increases in future temperatures, emissions and solar radiation that in turn lead to decreases in surface ozone."
Ozone and heat
Ozone pollution is not emitted directly. It forms as a result of chemical reactions that take place between nitrogen oxides and volatile organic compounds in the presence of sunlight.
These gases come from human activities such as combustion of coal and oil, as well as natural sources such as emissions from plants.
To examine the effects of climate change on ozone pollution, Pfister and colleagues looked at two scenarios.
In one, emissions of nitrogen oxides and volatile organic compounds from human activities would continue at current levels through 2050.
In the other, emissions would be cut by 60-70 percent. Both scenarios assumed continued greenhouse gas emissions with significant warming.
The researchers found that, if emissions continue at present-day rates, the number of eight-hour periods in which ozone would exceed 75 parts per billion (ppb) would jump by 70 percent on average across the United States by 2050.
The 75 ppb level over eight hours is the threshold that is considered unhealthy by the U.S. Environmental Protection Agency. (The agency is considering tightening the standard to a value between 65 and 70 ppb over eight hours.)
Overall, the study found that, 90 percent of the time, ozone levels would range from 30 to 87 ppb in 2050 compared with an estimated 31 to 79 ppb at present.
Although the range itself shifts only slightly, the result is a much larger number of days above the threshold considered unhealthy.
There are three primary reasons for the increase in ozone with climate change:
In the second scenario, Pfister and colleagues found that sharp reductions in nitrogen oxides and volatile organic compounds could reduce ozone pollution even as the climate warms.
In fact, 90 percent of the time, ozone levels would range from 27 to 55 ppb.
The number of instances when ozone pollution would exceed the 75 ppb level dropped to less than 1 percent of current cases.
"Our work confirms that reducing emissions of ozone precursors would have an enormous effect on the air we all breathe," Pfister said.
Pfister and a nationwide scientific team expect to learn more about the sources, chemistry and movement of air pollutants this summer when they launch a major field experiment known as FRAPPÉ along Colorado's Front Range.
The role of supercomputing
The study was among the first conducted on the new 1.5 petaflops Yellowstone supercomputer. The IBM system, operated by NCAR and supported by funding from NSF and the University of Wyoming, is one of the world's most powerful computers dedicated to research in the atmospheric and related sciences.
"High resolution models can consume significant time and resources on massive computers, but as shown in this research, they're often required for accurate regional ozone projections," said Irene Qualters, division director for Advanced Computing Infrastructure at NSF.
"Running these models wouldn't have been possible without the parallel processing power of the Yellowstone supercomputer, a critical part of NSF's cyberinfrastructure.
"The work will also help other researchers in related climate topics determine scenarios where coarse resolution is sufficient and, conversely, where high resolution is needed."
Thanks to its computing power, the scientists were able to simulate pollution levels hour-by-hour for 39 hypothetical summers.
This allowed the team to account for year-to-year variations in meteorological conditions, such as hot and dry vs. cool and wet, thereby getting a more detailed and statistically significant picture of future pollution levels.
To simulate the interplay of global climate with regional pollution conditions, the scientists turned to two of the world's leading atmospheric models, both based at NCAR and developed through collaborations in the atmospheric sciences community.
They used the Community Earth System Model, funded primarily by the U.S. Department of Energy and NSF, to simulate global climate as well as atmospheric chemistry conditions.
They also used an air chemistry version of the multiagency Weather Research and Forecasting Model to obtain a more detailed picture of regional ozone levels.
Even with Yellowstone's advanced computing speed, it took months to complete the complex simulations.
"This research would not have been possible even just a couple of years ago," said Pfister.
"Without the new computing power made possible by Yellowstone, you cannot depict the necessary detail of future changes in air chemistry over small areas, including the urban centers where most Americans live."
Cheryl Dybas, NSF-GEO, (703) 292-7734, email@example.com
Aaron Dubrow, NSF-CISE, (703) 292-4489, firstname.lastname@example.org
David Hosansky, NCAR, (303) 497-8611, email@example.com
Nanci Bompey, AGU, (202) 777-7524, firstname.lastname@example.org
NSF Grant: Collaborative Research: Developing a Next-Generation Approach to Regional Climate Prediction at High Resolution: http://www.nsf.gov/awardsearch/showAward?AWD_ID=1048829&HistoricalAwards=false
The National Science Foundation (NSF) is an independent federal agency that supports fundamental research and education across all fields of science and engineering. In fiscal year (FY) 2014, its budget is $7.2 billion. NSF funds reach all 50 states through grants to nearly 2,000 colleges, universities and other institutions. Each year, NSF receives about 50,000 competitive requests for funding, and makes about 11,500 new funding awards. NSF also awards about $593 million in professional and service contracts yearly.
Cheryl Dybas | Eurek Alert!
Field widens for environments, microbes that produce toxic form of mercury
12.10.2015 | DOE/Oak Ridge National Laboratory
Unexpected information about Earth's climate history from Yellow River sediment
09.10.2015 | Uppsala University
Physicists of TU Berlin and mathematicians of MATHEON are so successful that even the prestigious journal “Nature Communications” reported on their project.
Security in data transfer is an important issue, and not only since the NSA scandal. Sometimes, however, the need for speed conflicts to a certain degree with...
Having a light touch can make a hefty difference in how well animals and robots move across challenging granular surfaces such as snow, sand and leaf litter. Research reported October 9 in the journal Bioinspiration & Biomimetics shows how the design of appendages – whether legs or wheels – affects the ability of both robots and animals to cross weak and flowing surfaces.
Using an air fluidized bed trackway filled with poppy seeds or glass spheres, researchers at the Georgia Institute of Technology systematically varied the...
Nondestructive material testing (NDT) is a fast and effective way to analyze the quality of a product during the manufacturing process. Because defective materials can lead to malfunctioning finished products, NDT is an essential quality assurance measure, especially in the manufacture of safety-critical components such as automotive B-pillars. NDT examines the quality without damaging the component or modifying the surface of the material. At this year's Blechexpo trade fair in Stuttgart, Fraunhofer IZFP will have an exhibit that demonstrates the nondestructive testing of high-strength automotive body parts using 3MA. The measurement results are available in a matter of seconds.
To minimize vehicle weight and fuel consumption while providing the highest level of crash safety, automotive bodies are reinforced with elements made from...
The MICADO camera, a first light instrument for the European Extremely Large Telescope (E-ELT), has entered a new phase in the project: by agreeing to a Memorandum of Understanding, the partners in Germany, France, the Netherlands, Austria, and Italy, have all confirmed their participation. Following this milestone, the project's transition into its preliminary design phase was approved at a kick-off meeting held in Vienna. Two weeks earlier, on September 18, the consortium and the European Southern Observatory (ESO), which is building the telescope, have signed the corresponding collaboration agreement.
As the first dedicated camera for the E-ELT, MICADO will equip the giant telescope with a capability for diffraction-limited imaging at near-infrared...
Self-driving cars will be on our streets in the foreseeable future. In Graz, research is currently dedicated to an innovative driver assistance system that takes over control if there is a danger of collision. It was nature that inspired Dr Manfred Hartbauer from the Institute of Zoology at the University of Graz: in dangerous traffic situations, migratory locusts react around ten times faster than humans. Working together with an interdisciplinary team, Hartbauer is investigating an affordable collision detector that is equipped with artificial locust eyes and can recognise potential crashes in time, during both day and night.
Inspired by insects
01.10.2015 | Event News
30.09.2015 | Event News
17.09.2015 | Event News
13.10.2015 | Trade Fair News
13.10.2015 | Physics and Astronomy
13.10.2015 | Health and Medicine