Grand Challenges in Earthquake Engineering Research A Community Workshop Report
Committee for the Workshop on Grand Ch...
12 downloads
627 Views
604KB Size
Report
This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
Report copyright / DMCA form
Grand Challenges in Earthquake Engineering Research A Community Workshop Report
Committee for the Workshop on Grand Challenges in Earthquake Engineering Research— A Vision for NEES Experimental Facilities and Cyberinfrastructure Tools Committee on Seismology and Geodynamics Board on Earth Sciences and Resources Division on Earth and Life Studies
THE NATIONAL ACADEMIES PRESS╆
500 Fifth Street, N.W.╆
Washington, DC 20001
NOTICE: The project that is the subject of this report was approved by the Governing Board of the National Research Council, whose members are drawn from the councils of the National Academy of Sciences, the National Academy of Engineering, and the Institute of Medicine. The members of the committee responsible for the report were chosen for their special competences and with regard for appropriate balance. This study was supported by the National Science Foundation under contract No. CMMI-1047519. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation. International Standard Book Number-13: 978-0-309-21452-0 International Standard Book Number-10: 0-309-21452-1 Additional copies of this report are available from the National Academies Press, 500 Fifth Street, N.W., Lockbox 285, Washington, DC 20055; (800) 624-6242 or (202) 334-3313 (in the Washington metropolitan area); Internet: http://www.nap.edu. Cover: Cover design by Francesca Moghari. Copyright 2011 by the National Academy of Sciences. All rights reserved. Printed in the United States of America
The National Academy of Sciences is a private, nonprofit, self-perpetuating society of distinguished scholars engaged in scientific and engineering research, dedicated to the furtherance of science and technology and to their use for the general welfare. Upon the authority of the charter granted to it by the Congress in 1863, the Academy has a mandate that requires it to advise the federal government on scientific and technical matters. Dr. Ralph J. Cicerone is president of the National Academy of Sciences. The National Academy of Engineering was established in 1964, under the charter of the National Academy of Sciences, as a parallel organization of outstanding engineers. It is autonomous in its administration and in the selection of its members, sharing with the National Academy of Sciences the responsibility for advising the federal government. The National Academy of Engineering also sponsors engineering programs aimed at meeting national needs, encourages education and research, and recognizes the superior achievements of engineers. Dr. Charles M. Vest is president of the National Academy of Engineering. The Institute of Medicine was established in 1970 by the National Academy of Sciences to secure the services of eminent members of appropriate professions in the examination of policy matters pertaining to the health of the public. The Institute acts under the responsibility given to the National Academy of Sciences by its congressional charter to be an adviser to the federal government and, upon its own initiative, to identify issues of medical care, research, and education. Dr. Harvey V. Fineberg is president of the Institute of Medicine. The National Research Council was organized by the National Academy of Sciences in 1916 to associate the broad community of science and technology with the Academy’s purposes of furthering knowledge and advising the federal government. Functioning in accordance with general policies determined by the Academy, the Council has become the principal operating agency of both the National Academy of Sciences and the National Academy of Engineering in providing services to the government, the public, and the scientific and engineering communities. The Council is administered jointly by both Academies and the Institute of Medicine. Dr. Ralph J. Cicerone and Dr. Charles M. Vest are chair and vice chair, respectively, of the National Research Council. www.national-academies.org
COMMITTEE FOR THE WORKSHOP ON GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH—A VISION FOR NEES EXPERIMENTAL FACILITIES AND CYBERINFRASTRUCTURE TOOLS GREGORY L. FENVES, Co-chair, University of Texas at Austin CHRIS D. POLAND, Co-chair, Degenkolb Engineers, San Francisco, California ADAM J. CREWE, University of Bristol, United Kingdom RONALD T. EGUCHI, ImageCat, Inc., Long Beach, California JEROME F. HAJJAR, Northeastern University, Boston, Massachusetts JEROME P. LYNCH, University of Michigan, Ann Arbor MASAYOSHI NAKASHIMA, Kyoto University, Japan National Research Council Staff JASON ORTEGO, Project Manager and Research Associate DAVID A. FEARY, Project Director ERIC EDKIN, Senior Program Assistant
v
BOARD ON EARTH SCIENCES AND RESOURCES CORALE L. BRIERLEY, Chair, Brierley Consultancy, LLC, Highlands Ranch, Colorado KEITH C. CLARKE, University of California, Santa Barbara DAVID J. COWEN, University of South Carolina, Columbia WILLIAM E. DIETRICH, University of California, Berkeley ROGER M. DOWNS, Pennsylvania State University, University Park JEFF DOZIER, University of California, Santa Barbara WILLIAM L. GRAF, University of South Carolina, Columbia RUSSELL J. HEMLEY, Carnegie Institution of Washington, Washington, D.C. MURRAY W. HITZMAN, Colorado School of Mines, Golden EDWARD KAVAZANJIAN, JR., Arizona State University, Tempe ROBERT B. McMASTER, University of Minnesota, Minneapolis M. MEGHAN MILLER, UNAVCO, Inc., Boulder, Colorado ISABEL P. MONTAÑEZ, University of California, Davis CLAUDIA INÉS MORA, Los Alamos National Laboratory, New Mexico BRIJ M. MOUDGIL, University of Florida, Gainesville CLAYTON R. NICHOLS, Department of Energy, Idaho Operations Office (Retired), Ocean Park, Washington HENRY N. POLLACK, University of Michigan, Ann Arbor JOAQUIN RUIZ, University of Arizona, Tucson DAVID T. SANDWELL, Scripps Institution of Oceanography, La Jolla, California PETER M. SHEARER, University of California, San Diego REGINAL SPILLER, Frontera Resources Corporation (Retired), Houston, Texas RUSSELL E. STANDS-OVER-BULL, Anadarko Petroleum Corporation, Denver, Colorado TERRY C. WALLACE, JR., Los Alamos National Laboratory, New Mexico National Research Council Staff ANTHONY R. de SOUZA, Director ELIZABETH A. EIDE, Senior Program Officer DAVID A. FEARY, Senior Program Officer ANNE M. LINN, Senior Program Officer SAMMANTHA L. MAGSINO, Program Officer MARK D. LANGE, Associate Program Officer JENNIFER T. ESTEP, Financial and Administrative Associate NICHOLAS D. ROGERS, Financial and Research Associate COURTNEY R. GIBBS, Program Associate JASON R. ORTEGO, Research Associate ERIC J. EDKIN, Senior Program Assistant CHANDA IJAMES, Program Assistant
vi
COMMITTEE ON SEISMOLOGY AND GEODYNAMICS DAVID T. SANDWELL, Chair, Scripps Institution of Oceanography, La Jolla, California MICHAEL E. WYSESSION, Vice Chair, Washington University, St. Louis, Missouri J. RAMÓN ARROWSMITH, Arizona State University, Tempe EMILY E. BRODSKY, University of California, Santa Cruz JAMES L. DAVIS, Lamont-Doherty Earth Observatory, Palisades, New York STUART P. NISHENKO, Pacific Gas and Electric Company, San Francisco, California PETER L. OLSON, Johns Hopkins University, Baltimore, Maryland NANCY L. ROSS, Virginia Polytechnic Institute & State University, Blacksburg CHARLOTTE A. ROWE, Los Alamos National Laboratory, New Mexico BRIAN W. STUMP, Southern Methodist University, Dallas, Texas AARON A. VELASCO, University of Texas, El Paso
vii
Preface
Only three days after the Tohoku earthquake, the National Research Council convened a long-planned community workshop sponsored by the National Science Foundation. The purpose of the workshop was to give members of the earthquake engineering community an opportunity to identify the “Grand Challenges” for earthquake engineering research that would be needed to achieve an earthquakeresilient society. Based on these grand challenges, the participants were asked to identify the networked facilities, both experimental and cyberinfrastructure, needed to address the challenges. Six keynote speakers provided their ideas and perspectives about the grand challenges and key technologies that will contribute to earthquake resilience. Through discussions in breakout sessions and plenary sessions, the participants identified Grand Challenges in earthquake engineering and the general requirements for networked facilities to pursue Grand Challenge research on preparedness, mitigation, response, and recovery. The timing of the community workshop, held while the world was learning about the impact of the Tohoku earthquake, emphasized to the 53 participants that although great progress in earthquake resilience has been made over the past 40 years, the geological threats to the United States are substantial. Society must make a major commitment to increase the resilience of its communities, infrastructure, and citizens. The Grand Challenge problems and networked facilities identified by the workshop participants will accelerate the research needed for transformative solutions to achieve this goal.
Within a one-month period in 2011, two major earthquakes struck countries that are among the best prepared for earthquakes. The M6.1 South Island of New Zealand earthquake (February 22, 2011) had an epicenter 6 km from Christchurch. Less than three weeks later, the M9.0 Tohoku earthquake (March 11, 2011) hit the northeast region of Japan’s largest island, unleashing a massive tsunami. At the time of this publication, it is estimated that more than 25,000 people perished in the Tohoku earthquake and the economic toll in Japan may exceed 3 percent of its gross domestic product. These two earthquakes came on the heels of major earthquakes in Haiti and Chile within the past 18 months, providing stark reminders of the devastating impact major earthquakes have on the lives and economic stability of millions of people worldwide. The events in Haiti continue to show that poor planning and governance lead to long-term chaos. The steady recovery of Chile demonstrates that modern earthquake planning, proper construction, and mitigation activities facilitate rapid recovery. The recent earthquake in New Zealand underscores the importance of including resilience—the ability to recover quickly—as a goal of urban development, land use planning, and earthquake preparedness. Japan, as one of the most prepared nations, reminds us of the evolving nature of our understanding of major earthquakes and their consequences. There is uncertainty inherent in all aspects of earthquake engineering that needs to be addressed on an ongoing basis with transformative research, process and code development, and focused implementation programs.
Gregory L. Fenves, Co-chair Chris D. Poland, Co-chair
ix
Acknowledgments
tivity, evidence, and responsiveness to the study charge. The review comments and draft manuscript remain confidential to protect the integrity of the deliberative process. We wish to thank the following individuals for their participation in the review of this report:
The workshop would not have been successful without the contributions of those who attended; discussions were informative, professional, and conducted in a cooperative spirit. For providing excellent white papers and keynote presentations to the workshop, the committee would like to thank Gregory Deierlein, Reginald DesRoches, Omar Â�Ghattas, John Halloran, Laurie Johnson, and James Myers. The workshop was also greatly enhanced by the leadership of the breakout group moderators: John Egan, Ken Elwood, Kathleen Tierney, and Sharon Wood. Also to be acknowledged are the workshop participants: Richard Bowker, Mehmet Celebi, Raymond Daddazio, Leonardo Duenas-Osorio, Shirley Dyke, Ahmed Elgamal, Mohammed Ettouney, Charles Farrar, Kent Ferre, Steve French, Branko Glisic, Thomas Heaton, Jon Heintz, William Holmes, Muneo Hori, Kimberly Kurtis, Kincho Law, Bret Lizundia, Stephen Mahin, James Malley, Sami Masri, Peter May, Shamim Pakzad, Hope Seligson, Jonathan Stewart, Costas Synolakis, and Solomon Yim. All participants are listed in Appendix C of this document. This report has been reviewed in draft form by individuals chosen for their diverse perspectives and technical expertise, in accordance with procedures approved by the National Research Council’s (NRC’s) Report Review Committee. The purpose of this independent review is to provide candid and critical comments that will assist the institution in making its published report as sound as possible and to ensure that the report meets institutional standards for objec-
Richard Bowker, Memphis Light, Gas, and Water, Tennessee Russell A. Green, Virginia Polytechnic Institute and State University, Blacksburg Ronald O. Hamburger, Simpson Gumpertz & Heger, Inc., San Francisco, California Ann Patton, Ann Patton Company, LLC, Tulsa, Oklahoma Kathleen J. Tierney, University of Colorado at Boulder Although the reviewers listed above have provided many constructive comments and suggestions, they were not asked to endorse—nor did they see—the final draft of the workshop report before its release. The review of this report was overseen by Susan Hanson, Clark University, and William Holmes, Rutherford & Chekene. Appointed by the NRC, they were responsible for making certain that an independent examination of this report was carried out in accordance with institutional procedures and that all review comments were carefully considered. Responsibility for the final content of this report rests entirely with the authoring committee and the NRC.
x
Contents
OVERVIEW 1 1 INTRODUCTION Background, 5 Overview of the Workshop, 5 Organization of the Report, 8
5
2
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH Community Resilience Framework, 13 Decision Making, 14 Simulation, 15 Mitigation, 16 Design Tools, 18
11
3
NETWORKS OF FACILITIES Community Resilience Observatory, 22 Instrumented City, 23 Earthquake Engineering Simulation Center, 23 Earthquake Engineering Data Synthesis Center, 23 Earth Observation, 23 Rapid Monitoring Facility, 24 Sustainable Materials Facility, 24 Networked Geotechnical Centrifuges, 24 SSI Shaking Table, 24 Large-Scale Shaking Table, 24 Tsunami Wave Simulator, 24 Advanced Structural Subsystems Characterization Facility, 25 Non-Structural, Multi-Axis Testing Facility, 25 Mobile Facility for In Situ Structural Testing, 25
21
REFERENCES 27
xi
xii
CONTENTS
APPENDIXES A Breakout Session Presentations B White Papers Transformative Earthquake Engineering Research and Solutions for Achieving Earthquake-Resilient Communities, 44 Grand Challenges in Lifeline Earthquake Engineering Research, 51 Earthquake Engineering Research Needs in the Planning, Design, Construction, and Operation of Buildings, 58 Cyberinfrastructure-Intensive Approaches to Grand Challenges in Earthquake Engineering Research, 65 A Built Environment from Fossil Carbon, 70 Uncertainty Quantification and Exascale Computing: Opportunities and Challenges for Earthquake Engineering, 74 C Workshop Participants D Workshop Agenda E Committee Biographies
29 43
81 85 89
Overview
PURPOSE OF THE WORKSHOP
The committee organized the workshop into a series of keynote presentations, breakout sessions, and plenary sessions. Six keynote speakers were tasked with articulating, through their presentations and associated white papers (Appendix B), a vision that would help guide discussions among the workshop participants. Each speaker discussed a key component of earthquake engineering research— community, lifelines, buildings, information technology, materials, and modeling and simulation—and considered four cross-cutting dimensions—community resilience, preevent prediction and planning, design of infrastructure, and post-event response and recovery. Breakout sessions were the primary mechanism for brainstorming, analyzing, and documenting responses to the workshop questions outlined in the task. Four breakout sessions were structured along the cross-cutting dimensions, and one breakout session organized participants along disciplinary lines—buildings, lifelines, geotechnical/tsunamis, and community resilience. Each breakout session included a moderator, who served as the leader and chief spokesperson for the breakout group, and a committee member who served as rapporteur.
The George E. Brown, Jr. Network for Earthquake Engineering Simulation (NEES), supported by the National Science Foundation (NSF), is an important component of the National Earthquake Hazards Reductions Program (Â�NEHRP). NEHRP is a coordinated effort across four federal agencies to address earthquake risk in the United States. Since 2004, NEES researchers have produced significant advances in the science and technology for earthquake loss reduction that would not have been possible without the network’s experimental facilities and cyberinfrastructure. By Fiscal Year 2014, NSF will have supported 10 years of NEES operations and research. As part of NSF’s preparation of plans for Fiscal Year 2014 and beyond, NSF sought input from the broad earthquake engineering community on “Grand Challenges in Basic Earthquake Engineering Research,” with one consideration being that the program after 2014 need not be focused on—or limited to—existing facilities. At the request of NSF (see Statement of Task, Box S.1), the National Research Council (NRC) hosted a two-day workshop to give members of the community an opportunity to identify grand challenges and to describe networks of earthquake engineering experimental capabilities and cyberinfrastructure tools that could contribute to addressing these challenges.
SUMMARY OF WORKSHOP DISCUSSIONS This report summarizes the major points and ideas expressed during the workshop. It is not intended to be a comprehensive summary of all topics and issues relevant to earthquake engineering research. The observations or views contained in this report are those of individual participants and do not necessarily represent the views of all workshop participants, the committee, or the NRC. Therefore, references in the report to workshop “participants” do not imply that all participants were polled or that they necessarily agreed with the particular statements. In addition, the grand challenge problems and networked facilities discussed in the following sections were suggested by breakout group participants and they do not represent conclusions or recommendations of the committee or the NRC.
WORKSHOP PLANNING An NRC steering committee was established to organize the workshop, which was held on March 14–15, 2011, at the NRC’s Beckman Center in Irvine, California. Workshop participants included 37 researchers and practitioners, drawn from a wide range of disciplines, to focus on the two key questions in the task statement. In addition, observers from NSF, NSF contractors, NEHRP, and the current NEES OperaÂ�tions Center attended the discussions. Altogether, there were 52 workshop attendees, including the committee and NRC staff (Appendix C). 1
2
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
BOX S.1 Statement of Task The National Science Foundation (NSF) supports the George E. Brown, Jr. Network for Earthquake Engineering Simulation (NEES), as a component of the National Earthquake Hazards Reduction Program (NEHRP). In Fiscal Year (FY) 2014, NSF will have supported 10 years of NEES operations and research, and seeks an evaluation of next-generation U.S. needs for earthquake engineering research beyond 2014. A National Research Council committee will organize a public workshop on the Grand Challenges for earthquake engineering research, to bring together experts to focus on two questions: 1. What are the high-priority Grand Challenges in basic earthquake engineering research that require a network of earthquake engineering experimental facilities and cyberinfrastructure? 2. What networked earthquake engineering experimental capabilities and cyberinfrastructure tools are required to address these Grand Challenges? The workshop will feature invited presentations and discussion. The committee will develop the agenda, select and invite speakers and discussants, and moderate the discussion. Workshop participants will be asked to describe the experimental infrastructure capabilities and cyberinfrastructure tools in terms of requirements, rather than by reference to any existing or specifically located future facilities. In responding to the foregoing questions, workshop participants will also be asked to consider future technical and conceptual advances with the potential to influence future earthquake hazard research, such as early warning systems, new materials, sustainability, high-performance computing and networking, modeling, sensor and monitoring technologies, and other factors identified by the committee. The committee will prepare a report summarizing discussions at the workshop; the report will not include findings or recommendations.
Grand Challenges in Earthquake Engineering Research Grand challenges in earthquake research are the problems, barriers, and bottlenecks in the earthquake engineering field that hinder realization of the NEHRP vision—“A nation that is earthquake resilient in public safety, economic strength, and national security” (NEHRP, 2008). As such, they define frontiers in basic earthquake engineering research that would be needed to provide transformative solutions for achieving an earthquake-resilient society. Thirteen grand challenge problems emerged over the course of the workshop. The committee has summarized them in terms of five overarching Grand Challenges, Â�described below, in order to capture interrelationships and crossovers among the 13 problems and to highlight the interdisciplinary nature of their potential solutions. Participants noted that grand challenge problems do not stand alone; they are complex, and this complexity exists not only within earthquake engineering but also in earthquake engineering’s position among other competing social challenges. As such, addressing a grand challenge problem involves consideration of a variety of barriers—economic, regulatory, policy, societal, and professional—along with the scientific and technological solutions. The five overarching Grand Challenges are intended to serve as useful focal points for discussions among stakeholders and decision makers planning future investment toward achieving a more earthquake-resilient nation.
1. Community Resilience Framework: A common theme noted by participants was that the earthquake engineering community currently lacks an interactive and comprehensive framework for measuring, monitoring, and evaluating community resilience. Such a framework could apply innovative methodologies, models, and data to measure community performance at various scales, build on the experience and lessons of past events, and help ensure that past and future advances in building, lifelines, urban design, technology, and socioeconomic research result in improved community resilience. Such a framework also could advance our understanding of both the direct and indirect impacts of earthquakes so that community-level interactions and impacts can be better characterized. 2. Decision Making: Another sentiment reiterated during the workshop was that current research findings related to community resilience do not adequately influence decisions and actions on the part of key decision makers, such as privatesector facility owners and public-sector institutions. Communities typically build based on traditional standards, and when affected by major earthquakes, they respond and recover based on intuition, improvisation, and adaptive behaviors
3
OVERVIEW
that are drawn from the individuals available to participate. Â�Consequently, the lessons learned in one community and event rarely translate to the next comÂ�munity affected. Participants suggested that achieving earthquake resilience could involve a community-based, holistic approach that includes decisions and actions that are based on overarching goals, a clear understanding of the built environment, rapid and informed assessment data, and planned reconstruction and recovery. Mechanisms for motivating action could include developing incentives to promote community development and pre-event planning; simulationbased decision-making strategies for use in community development, pre-event planning, in early response post event, and through the long-term recovery process; state-of-the-art decision-making tools that will lead to more efficient resource allocations; and methodologies and tools that allow decision makers to compare different strategies for post-earthquake reconstruction and long-term pre-earthquake Â�mitigation. 3. Simulation: Participants noted that knowledge of the inventory of infrastructure components and points of connection between different infrastructure types is lacking within the earthquake engineering community. They identified a need for scalable tools that autonomously create an accurate database of all infrastructure components, including points of interdependency with other infrastructure components. Empowered with this complete mapping of an urban region’s infrastructure systems, powerful simulation technologies could model the time and spatial impacts of a seismic event at all length scales spanning from the component scale to the regional scale, and from disaster response to community recovery. 4. Mitigation: A large earthquake or tsunami in a highly populated region of the United States would cause massive damage to the built environment and communities in the region, and the resulting social and economic consequences would cascade across the country, particularly if major energy, transportation, or supply hubs are affected. Key characteristics of this Grand Challenge include developing strategies to measure, monitor, and model community vulnerability, motivations, and mitigation strategies, and establishing mitigation solutions for the community’s most vulnerable sectors. Participants suggested that mitigation solutions could be based on the use of a new generation of simulation tools and design solutions coupled with up-to-date information available from distributed sensing systems. Development of better approaches for renewal and
retrofit of the built environment’s most vulnerable sectors would help ensure a safer environment and a more resilient community. 5. Design Tools: Participants suggested that developing and exploiting new emerging materials and innovative structural concepts and integrating them within design tools could dramatically improve the performance of all types of infrastructure and increase earthquake resilience in ways that are also sustainable. There is a wide range of sustainable, highly resilient, new materials that can offer opportunities to significantly change the way infrastructure is designed and constructed. Harnessing the power of performance-based earthquake engineering could achieve a resilient infrastructure that incorporates these innovative new materials and structural systems. Networks of Facilities The second goal of the workshop was for participants to identify the general requirements for networked earthquake engineering experimental capabilities and cyberÂ�infrastructure tools associated with addressing the grand challenge problems. The suggested experimental facilities cover testing and monitoring over a wide range of scales, loading regimes, boundary conditions, and rates on laboratory and field (in situ) specimens. Cyberinfrastructure tools are also important for capturing, analyzing, and visualizing experiments and for supporting the advanced simulations discussed in the workshop. Participants described 14 facilities that could contribute to solving the grand challenge problems: 1. Community resilience observatory: Such an observatory could encompass interlinked facilities that function as a laboratory without walls, integrating experimental testing and simulations with a holistic understanding of communities, stakeholders, decisions, and motivations. 2. Instrumented city: An instrumented testbed in a high-risk, urban environment could provide invaluable data about the performance of the community and allow unprecedented research on studying decision-making processes for development and calibration of comprehensive, community models. 3. Earthquake engineering simulation center: Such a center could bring together earthquake engineering researchers with experts in algorithm development, computational and statistical methods, and high-end computational and cloud development methodologies to enable transformative advances in modeling and simulation.
4
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
4. Earthquake engineering data synthesis center: Such a center could offer the research community a large-scale database system for ingesting data sources from a variety of sensor types including imaging, remote sensing, video, and information management systems. 5. Earth observation: Earth observation systems could provide an integration of continuous and multi-sensor (e.g., aerial, satellite, and unmanned aerial vehicle) observations of communities at various scales for the purpose of characterizing the physical attributes of communities and monitoring the effects of earthquakes (e.g., damage assessment and recovery). 6. Rapid monitoring facility: Such a facility could provide the earthquake engineering community with a suite of highly portable sensing and data acquisition tools that could be rapidly deployed to structures, geo-facilities, and lifelines to monitor their stability after seismic events. 7. Sustainable materials facility: Partnering with material science facilities could lead to the development and testing of new construction grade �materials that are self-healing, capable of energy capture, or ultra-high strength, and to understand the use of sustainable materials for earthquake engineering applications. A sustainable materials facility could test these materials under the conditions they may experience when used in construction account�ing for the influence of aging and �degradation. 8. Networked geotechnical centrifuges: Networked geotechnical centrifuges, each including innovative capabilities for robotic manipulation and actuation within the centrifuge container during the experiment, could allow new types of experimental modeling of landslides (including submarine landslides), liquefaction, and tsunamis. 9. SSI shaking table: A large-scale, dynamic shaking table designed for soil-structure interaction (SSI)
experiments could enable a significant throughput of SSI experiments to help advance knowledge of this crucial component of earthquake engineering. 10. Large-scale shaking table: Testing complete structures or full-scale subsystems in multiple directions could provide fundamental knowledge for understanding the response of actual construction and the contributions of lateral and gravity load resisting systems and non-structural systems, validating post-earthquake evaluation methods for damaged structures. 11. Tsunami wave simulator: Such a revolutionary new facility could combine a tsunami wave basin with the capability to shake the ground to simulate liquefaction and subsidence. 12. Advanced structural subsystems characterization facility: Such a facility could test full-sized or closeto-full-scale subsystems and components �under fully realistic boundary and loading conditions, to replicate the effects of corrosion, accelerated aging, and fatigue, and have the capability for multi-axial loading, high-temperature testing, and high pressures. It could enable the development of more accu�rate structural models needed for characterization of subsystems, components, and materials. 13. Non-structural, multi-axis testing facility: A high-performance multi-axis facility could be developed with the frequency range and levels of �motion to investigate and characterize the performance of non-structural elements (e.g., partitions) and other content (e.g., shelving, information technology equipment, lighting, electrical and �mechanical equipment) in three dimensions within a building or other infrastructure. 14. Mobile facility for in situ structural testing: A suite of highly portable testing equipment in such a facility could include shakers, actuators, sensors, and high-resolution data acquisition systems that could enable structures, lifelines, or geotechnical systems to be tested in place.
1 Introduction
OVERVIEW OF THE WORKSHOP
The National Science Foundation (NSF) supports the George E. Brown, Jr. Network for Earthquake Engineering Simulation (NEES), an important component of the National Earthquake Hazards Reduction Program (NEHRP). Since 2004, NEES researchers have produced significant advances in the science and technology for earthquake loss reduction that would have not been possible without the network’s experimental facilities and cyberinfrastructure. By Fiscal Year (FY) 2014, NSF will have supported 10 years of NEES operation and research. Looking beyond 2014, NSF asked the National Research Council (NRC) to conduct a community workshop to describe the Grand Challenges for earthquake engineering research related to achieving earthquake resilience. This report summarizes the discussions at the workshop.
Planning Committee A steering committee was established by the NRC to organize the workshop and to write a report summarizing what transpired at the workshop. Committee members were selected for their expertise in earthquake engineering research, broadly defined, with a focus on the use of experimental facilities and the application of cyberinfrastructure to engineering research. Special effort was made to involve members who were not strongly associated with existing NEES facilities so that the workshop could take a fresh look at the Grand Challenges and facilities requirements beyond 2014. The committee met in December 2010 to plan the workshop and again immediately following the workshop to organize workshop outputs into a report.
BACKGROUND NEHRP is a multi-agency program focused on reducing losses due to earthquakes. It includes programs at the Â�National Institute of Standards and Technology, Federal Emergency Management Agency, NSF, and U.S. Geological Survey. A major component of NSF’s role in NEHRP is focused on NEES, a large-scale investment in a nationally distributed network of shared engineering facilities for experimental and computational research—a national infrastructure for testing geoÂ�technical, structural, and nonstructural systems (see Box 1.1 for a description of the existing NEES network). With the current NEES network scheduled to end in 2014, NSF has sought community input for the preparation of plans, for FY 2014 and beyond, to address Grand Challenges in basic earthquake engineering research. NSF has stipulated that future investments in networked earthquake engineering research infrastructure beyond 2014 should not be focused on—or limited to—existing facilities but would build on the synergies provided by networked facilities and cyberinfrastructure tools to achieve solutions to the grand challenge problems.
Approach for the Workshop In accordance with the Statement of Task, the committee designed the workshop to look beyond 2014 and focus on two key questions: • What are the high-priority Grand Challenges in basic earthquake engineering research that require a network of earthquake engineering experimental facilities and/or cyberinfrastructure?1 • What are the general requirements for experimental facilities and cyberinfrastructure that will be needed to most effectively address the identified Grand Challenges?
1 The committee understood the first question of “networks of facilities and cyberinfrastructure,” not to require both but to allow a network of experimental facilities or a network of cyberinfrastructure services. Consequently, the word “and”—as written in the task statement (see Box S.1)—was interpreted as “and/or.”
5
6
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
BOX 1.1 The Existing NEES Network The primary focus of NEES is on the research community and practicing engineers who develop the innovations necessary to reduce the impact of seismic disasters. The NEES network infrastructure encompasses management headquarters; 14 earthquake engineering and tsunami research facility sites located at universities across the United States (available for testing on-site, in the field, or remotely); and cyberinfrastructure operations that connect the work of the experimental facilities, researchers, educators, and students. SOURCE: http://nees.org/.
The committee suggested that the Grand Challenges would define the frontiers in basic earthquake engineering research needed to provide transformative solutions for achieving an earthquake-resilient society. Transformative solutions to the Grand Challenges could be achieved by improved design codes, public policies, innovative systems, design and analysis methods, and sensing and actuation technologies embedded in the built environment. Workshop participants were asked to address the key questions without regard to the current capabilities or limitations of the existing NEES facilities. Paraphrasing the National Academy of Engineering’s (NAE) Grand Challenges for Engineering, a grand challenge is a large and complex problem that needs to be mastered to ensure the sustainability of civilization and the health of its citizens while reducing individual and societal vulnerabilities (NAE, 2008). A grand challenge will not be met without finding ways to overcome the barriers that block its accomplishment. The NEHRP vision—“A nation that is earthquake resilient in public safety, economic strength, and national security” (NEHRP, 2008)—is a grand challenge by the NAE definition. A fundamental goal of this workshop, therefore, was to describe the earthquake engineering challenges in terms of problems, barriers, and bottlenecks that must be solved to realize the NEHRP vision. Workshop Organization The workshop was held on March 14–15, 2011, at the NRC’s Beckman Center in Irvine, California. Workshop participants included 37 researchers and practitioners drawn from a wide range of disciplines to focus on the two key questions in the task statement. In addition, seven observers from NSF and the broader earthquake engineering research community attended the discussions. Altogether, there were 53 workshop attendees, including the committee and NRC staff. The committee invited six keynote speakers to the workshop to inform discussions about the Grand Challenges and rapid advances in technology. Through their presentations
and associated white papers (see Appendix B), which were distributed prior to the workshop, the speakers were tasked with articulating a vision that would help guide workshop discussions. The first three keynote speakers—Laurie Â�Johnson, Laurie Johnson Consulting; Reginald DesRoches, Georgia Institute of Technology; and Gregory Deierlein, Stanford University—presented their ideas for transformative earthquake engineering research in the categories of community, lifelines, and buildings, respectively. Each considered four dimensions: community resilience, preevent prediction and planning, design of infrastructure, and post-event response and recovery (see Box 1.2). To facilitate discussion on the advances in technology, three additional keynote speakers—James Myers, Rensselaer Polytechnic Institute; John Halloran, University of Michigan; and Omar Ghattas, University of Texas at Austin—presented roadmaps for information technology, materials, and modeling and simulation, respectively. The technology keynote speakers introduced the workshop participants to the transformative possibilities of technology for earthquake engineering beyond 2014. Additionally, two workshop participants—Ken Â�Elwood, University of British Columbia, and Thomas Heaton, Â�California Institute of Technology—provided their observations on the two recent devastating earthquakes in New Â�Zealand and Japan (see Box 1.3). Gregory Fenves, co-chair of the committee, spoke briefly on behalf of Masayoshi Â�Nakashima, committee member, who was unable to attend the workshop because of the major earthquake in Japan which had occurred just days before the workshop. Breakout sessions were the primary mechanism for brainstorming, analyzing, and documenting responses to the two key workshop questions. Four breakout sessions were structured along the dimensions described in Box 1.2, and one breakout session organized participants along disciplinary lines: buildings, lifelines, geotechnical/tsunamis, and community resilience. Each breakout session included a moderator, who served as both leader of the breakout session and facilitator of open and organized discussion, and a committee member who served as rapporteur. The Â�moderators—
7
INTRODUCTION
BOX 1.2 Four Dimensions as an Organizing Principle for the Workshop As an organizing principle for the workshop, the committee defined four dimensions to achieve the vision for an earthquake-resilient society. Examples of topics in each dimension are defined below. These were used as a starting point for the discussions at the workshop. Community resilience • Defining community response and recovery needs. • Obtaining community-based information and experiences that can be used for policy development. • Collecting, processing, analyzing, and disseminating information. • Pervasive information sharing and decision making through social networking and crowd-sourcing technology. • Understanding social dynamics that influence community decisions and actions. Pre-event prediction and planning • Damage prediction and the estimation of the impacts and losses for individual buildings, lifelines, and societal systems. • Validated and reliable models of soil-foundation-building systems, non-structural systems, and building contents. • Design of lifeline systems for multiple performance objectives and multiple levels of ground motion. • Models of inventory that can be validated and updated for regional impact assessment and loss estimation. • Predictive model of system performance and interdependencies. • Social, human, and economic resilience modeling, and the effects on adaptability after a disaster. • Modeling of effects of governance on resilience, such as regulatory regimes, emergency decision-making processes, and recovery policies. Design of infrastructure • Analysis and design approaches, strategies, and methods for systems, components, and materials, including new infrastructure, rehabilitation of existing infrastructure, and repair of damaged infrastructure. • Infrastructure design for individual buildings, lifelines, and urban environments as complex systems. • Transparent and performance-based approaches for buildings and lifelines along with other approaches that achieve multiple objectives for resiliency. Post-event response and recovery • Post-event sensing, damage diagnosis, and prognosis of individual facilities and interdependent infrastructure systems in dense urban environments. • Use of sensing systems for emergency response, including assessment, prioritization, dispatching, and decision making. • Real-time model updating and validation. • Social networking and crowd-sourcing technologies for understanding complex societal dynamics, including temporary changes in governance after an event and during recovery.
Kathleen Tierney, University of Colorado, Boulder; John Egan, AMEC Geomatrix; Ken Elwood, University of British Columbia; and Sharon Wood, University of Texas at Austin— also served as chief spokespersons for their breakout group in plenary sessions. Each breakout session allowed Â� ample time for discussion, interaction, and iteration, followed by a report in the plenary sessions with refinement by participants. Summary of Keynote Presentations The first three presentations focused on identifying and describing the Grand Challenges. Laurie Johnson discussed needs and opportunities for networked facilities and cyberÂ� infrastructure in support of basic and applied research on
community resilience, including the need to develop more robust models of building risk/resiliency and aggregate inventories of community risk/resiliency for use in mitigation, land use planning, and emergency planning. Reginald DesRoches discussed a number of challenges faced by lifeline facilities, including their wide range in scale and spatial distribution, the fact that lifelines are partially or completely buried and are therefore strongly influenced by soil-structure interaction, their increasing interconnectedness, and their aging and deterioration. Gregory Deierlein discussed methods for addressing the research needs and challenges for buildings, which he distinguished between those associated with either pre-earthquake planning, design and construction, or post-earthquake response, evaluation, and restoration.
8
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
BOX 1.3 Observations from New Zealand and Japan Ken Elwood, University of British Columbia, presented his observations from the February 22, 2011, earthquake in Christchurch, New Zealand, as three lessons relevant to the Grand Challenges in earthquake engineering. They are summarized below: 1. Aftershocks a. Consideration of aftershocks (e.g., residual capacity) should be incorporated into performance-based seismic design. b. We need a better understanding of the seismology of aftershocks and their broader effects within the seismological environment, and to incorporate this knowledge into post-earthquake decisions and recovery strategies. 2. Influence of soil-structure interaction (including liquefaction) on structural response a. Liquefaction can both help and hinder structural performance (damage to the building may occur, but the contents of the building will remain intact). b. A more holistic approach to buildings is needed (e.g., one that considers geotechnical systems). c. Large-scale studies are needed to study the interaction between the building and the foundation. 3. Community resilience a. Damage to the community extends far beyond lives lost; destruction of historical buildings and landmarks can have huge impacts on a city’s character and identity. b. A major public policy challenge is how to protect existing buildings (especially very old buildings with little real estate value). c. We need to evaluate the impact that post-earthquake assessment has on the ability for the city to recover. The following are comments from committee member Masayoshi Nakashima, Kyoto University, regarding the 2011 Tohoku Earthquake. These were based on preliminary information available in Japan on March 14, 2011. 1. An extremely large rupture of more than 400 km was not anticipated by seismologists. 2. A huge tsunami caused complete devastation of many towns and villages and large loss of life. Damage and deaths from tsunamis appear to be much greater than from the earthquake shaking. 3. There was significant subsidence (about 1 to 2 meters) of coast lines, which is speculated to have aggravated the tsunami damage. 4. Urban damage, such as observed in Sendai, is particularly characterized by the loss of lifelines. 5. The performance of hundreds of high-rises and base-isolated buildings in the Tokyo metropolitan area appears to be good. Many days are needed to collect associated data. 6. There is widespread disruption in the Tokyo metropolitan area, because of a shortage of electric power. 7. Post-earthquake responses of the central and local governments are being tested. Several hundred thousand people were forced to move to evacuation centers. 8. Technical and social response to nuclear accidents is a major issue for the country. 9. The earthquake caused large fires, including at oil tank farms. 10. There is severe liquefaction in areas of reclaimed land.
The second three presentations focused on advances in technology. James Myers discussed information technology and explored the potential for increased computing power, data sizes, and sensor density—combined with a rapidly increasing capability to focus those resources on demand and to automate larger and more complex tasks—to further progress on the Grand Challenges. John Halloran discussed new materials and proposed designing a built environment with more resilient, lighter, stronger, and more sustainable materials based on fossil carbon. Finally, Omar Ghattas discussed opportunities to extend large-scale simulation-based seismic hazard and risk analysis from its current reliance
on deterministic earthquake simulations to those based on stochastic models.
ORGANIZATION OF THE REPORT This report is the committee’s summary of what transpired at the workshop. It reflects only those topics addressed in workshop presentations, discussions, and background Â�papers, and it is not intended to be a comprehensive summary of all topics and issues relevant to earthquake engineering research. The observations or views contained in this report are those of individual participants and do not necessarily rep-
INTRODUCTION
resent the views of all workshop participants, the committee, or the NRC. Therefore, references in the report to workshop “participants” do not imply that all participants were polled or that they necessarily agreed with the particular statements. In addition, the grand challenge problems and networked facilities discussed in the following sections were suggested by breakout group participants and they do not represent conclusions or recommendations of the committee or the NRC. Chapter 2 describes the Grand Challenges in earth-
9 quake engineering research that require a network of earthquake engi�neering experimental facilities and cyber� infrastructure, and Chapter 3 summarizes those requirements. Appendix A contains the final breakout group presentations, and Appendix B contains the six white papers presented by the keynote speakers at the workshop. A list of workshop participants and the agenda are given in �Appendixes C and D, respectively. Appendix E presents biographical sketches of the committee members.
2 Grand Challenges in Earthquake Engineering Research
A fundamental goal of the workshop was to describe the high-priority grand challenges in earthquake engineering research, which are represented by the problems, barriers, and bottlenecks in the earthquake engineering field that hinder realization of the National Earthquake Hazards Reduction Program (NEHRP) vision. Thirteen grand challenge problems emerged over the course of the workshop. The committee has summarized them in terms of five overarching Grand Challenges to capture interrelationships and crossovers among the 13 problems and to highlight the interdisciplinary nature of their potential solutions. Participants noted that grand challenge problems do not stand alone; they are complex, and this complexity exists not only within earthquake engineering but also in earthquake engineering’s position among other competing social challenges. As such, addressing a grand challenge problem involves consideration of a Â�variety of barriers—economic, regulatory, policy, societal, and Â�professional—along with the scientific and technological solutions. The five overarching Grand Challenges are intended to serve as useful focal points for discussions among stakeholders and decision makers planning future investment. Table 2.1 shows the grouping of the 13 problems into the five overarching Grand Challenges, and it also maps each grand challenge problem to the disciplinary breakout group from which it originated. These Grand Challenges are:
framework also could advance understanding of both the direct and indirect impacts of earthquakes so that community-level interactions and impacts can be �better characterized. 2. Decision Making: Another sentiment reiterated during the workshop was that current research findings related to community resilience do not adequately influence decisions and actions on the part of key decision makers, such as private-sector facility owners and public-sector institutions. Communities typically build based on traditional standards, and when affected by major earthquakes, they respond and recover based on intuition, improvisation, and adaptive behaviors that are drawn from the individuals available to participate. Consequently, the lessons learned in one community and event rarely translate to the next community affected. Participants suggested that achieving earthquake resilience could involve a community-based, holistic approach that includes decisions and actions that are based on overarching goals, a clear understanding of the built environment, rapid and informed assessment data, and planned reconstruction and recovery. Mechanisms for motivating action could include developing incentives to promote community development and pre-event planning; simulation-based decision-making strategies for use in community development, pre-event planning, early response post event, and through the long-term recovery process; state-of-the-art decisionmaking tools that will lead to more efficient resource allocations; and methodologies and tools that allow decision makers to compare different strategies for post-earthquake reconstruction and long-term preearthquake mitigation.
1. Community Resilience Framework: A common theme noted by workshop participants was that the earthquake engineering community currently lacks an interactive and comprehensive framework for measuring, monitoring, and evaluating community resilience. Such a framework could apply innovative methodologies, models, and data to measure community performance at various scales, build on the experience and lessons of past events, and ensure that past and future advances in building, lifelines, urban design, technology, and socioeconomic research result in improved community resilience. Such a
3. Simulation: Participants noted that knowledge of the inventory of infrastructure components and points 11
12
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
TABLE 2.1╇ Grouping of 13 Grand Challenge Problems into the Five Overarching Grand Challenges. OVERARCHING GRAND CHALLENGES Community Resilience Framework
Decision Making
Simulation
Mitigation
Design Tools
â•⁄ 1. Framework for Measuring, Monitoring, and Evaluating Community Resilience
√
√
√
√
√
â•⁄ 2. Motivating Action to Enhance Community Resilience
√
√
√
√
â•⁄ 3. Develop a National Built Environment Inventory
√
√
√
√
â•⁄ 4. Multi-Scale Seismic Simulation of the Built Environment
√
√
√
â•⁄ 5. Integrated Seismic Decision Support
√
√
√
â•⁄ 6. Risk Assessment and Mitigation of Vulnerable Infrastructure
√
√
√
√
â•⁄ 7. Protect Coastal Communities
√
√
√
√
â•⁄ 8. Regional Disaster Simulator
√
√
√
√
â•⁄ 9. High Fidelity Simulation
√
√
√
10. New Sustainable Materials and Systems for Earthquake Resilience
√
√
√
√
11. Harnessing the Power of Performance Based Earthquake Engineering (PBEE) to Achieve Resilient Communities
√
√
√
√
12. Rapid Post-Earthquake Assessment
√
√
√
13. Reconstruction and Recovery
√
√
√
Dimension (Breakout Group)
Grand Challenge Problem
Community Resilience
Pre-event Prediction and Planning
Design of Infrastructure
Post-event Response and Recovery
√
√
NOTE: The dimension column on the left maps each grand challenge problem to the breakout group from which it originated; note that the grand challenge problems do not represent consensus views of the breakout groups, but rather suggestions by individuals or groups of individuals during the breakout group discussions (see Appendix A).
of connection between different infrastructure types is lacking within the earthquake engineering community. They identified a need for scalable tools that autonomously create an accurate database of all infrastructure components, including points of interdependency with other infrastructure components. Empowered with this complete mapping of an urban center’s infrastructure systems, powerful simulation technologies could model the time and spatial Â�impacts of a seismic event at all length scales spanning from the component scale to the regional scale, and from disaster response to community recovery. 4. Mitigation: A large earthquake or tsunami in a highly populated region of the United States would cause massive damage to the built environment and communities in the region, and the resulting social and economic consequences would cascade across the country, particularly if major energy, transportation, or supply hubs are affected. Key characteristics of this Grand Challenge include developing strategies to measure, monitor, and model community vulner-
ability, motivations, and mitigation strategies, and establishing mitigation solutions for the community’s most vulnerable sectors. Participants suggested that mitigation solutions could be based on the use of a new generation of simulation tools and design solutions coupled with up-to-date information available from distributed sensing systems. Development of better approaches for renewal and retrofit of the built environment’s most vulnerable sectors would help ensure a safer environment and a more resilient community. 5. Design Tools: Participants suggested that developing and exploiting new emerging materials and innovative structural concepts and integrating them within design tools could dramatically improve the performance of all types of infrastructure and increase earthquake resilience in ways that are also sustainable. There is a wide range of sustainable highly resilient, new materials that can offer opportunities to significantly change the way infrastructure is designed and constructed. Harnessing the power of
13
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
performance-based earthquake engineering (PBEE) could achieve a resilient infrastructure that incorporates these innovative new materials and structural systems. The five overarching Grand Challenges are summarized in the sections that follow. Characteristics of each Grand Challenge are given, along with transformative approaches to solving the grand challenge problems and the potential resulting impacts. Appendix A contains the original descriptions of the 13 grand challenge problems from the breakout sessions.
COMMUNITY RESILIENCE FRAMEWORK Description of the Problem Participants noted that although research has yielded numerous findings related to community resilience, many of these findings do not influence decisions or actions by key decision makers including private-sector facility owners and public-sector institutions.1 Characterizing the interactions and impacts at a community level necessitates an understanding of both the direct and indirect impacts of earthquakes, and a framework for measuring, monitoring, and evaluating community resilience could help ensure that past and future advances in building, lifelines, urban design, technology, and socioeconomic research result in improved community resilience. Such a framework could Â�apply innoÂ�vative methodologies, models, and data to measure community performance at various scales—e.g., building, lifeline, and community—and build on the experience and lessons of past events. Participants reiterated that such an interactive and comprehensive framework is lacking within the earthquake engineering community. In addition, many participants noted a need for basic research on the different mechanisms for motivating action. This includes information that stakeholders may use to quantify the costs and benefits of various mitigation strategies and the incentives for action that are meaningful to various constituencies, ranging from laws and regulations to informally applied norms. Characteristics of the Grand Challenge Because resilience is multi-dimensional and multi-scale, achieving resilience requires a multi-disciplinary approach. The earthquake engineering research community is, for example, unable at this time to define and measure multiple dimensions of resilience. Workshop participants discussed the need for a characterization of resiliency in terms of scale and metrics that are both applicable for diverse systems and for their interdependencies. Because researchers do not have 1 See the white paper in Appendix B by Laurie Johnson, the keynote speaker on community resilience: “Transformative Earthquake Engineering Research and Solutions for Achieving Earthquake-Resilient Communities.”
standard methods or measures for resiliency, it is difficult to determine when resiliency has been achieved. This is because current engineering approaches are limited in their ability to characterize resilience outcomes or to characterize them in ways that are meaningful for end users. Transformative Approaches to the Solution Many workshop participants emphasized that characterizing community resilience will require a significant shift in how the performance of communities is quantified. For example, existing research programs in earthquake engineering mainly focus on the performance of individual components or systems (e.g., buildings and specific lifeline systems), whereas understanding the performance of a community requires an understanding of the interactions among all of these components. Many questions still exist, including: how does the performance of an electric power system affect the performance of other lifeline systems? How does the disruption of power affect local and regional businesses? How does an industry in an affected region impact other industries that may not have been directly impacted by damage? Multi-scale modeling of resilience could effectively relate these diverse interactions. Another issue that has impeded the ability to measure and understand community resilience is the lack of historical data on recovery of communities from past disasters. Participants discussed the potential for a national observatory network to address the disaster vulnerability and resilience of communities using methodologies applied consistently over time and space, with attention to complex interactions between changes in social systems, the built environment, and the natural environment. They cited a 2008 workshop sponsored by the U.S. Geological Survey and the National Science Foundation that discussed the structure of such a network, called the Resiliency and Â�Vulnerability Â�Observatory Network (Peacock et al., 2008). Output from this network could help foster many research projects on community resilience including: • Developing and testing community resiliency metrics at different scales (e.g., communities, regions) and for different community components (e.g., buildings, lifelines, social networks, economy). • Researching, developing, and testing various methods for quantifying resilience and determining the best method for stakeholder decision making. • Creating a resilience observation pilot study, which could be a candidate city, neighborhood, or group of buildings (see “Instrumented City”), setting a baseline, and observing actions/changes over time to define metrics and timeframes of resiliency dynamics. • Encouraging the development of quantitative recovery models and developing theoretically and empirically based models of post-earthquake Â�recovery
14
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
processes. For example, participants noted that Â�models could be integrated across dimensions of recovery—infrastructures, Â� housing, business/Â� commercial facilities, public institutions, social/Â� economic processes—and incorporated into simulation models that forecast recovery rates and patterns after major earthquakes. These models could be designed to consider resilience, adaptation, sustainability, and mitigation. • Developing multi-scale simulation models that link the performance of buildings and lifelines to communities. • Developing data-intensive methods for using public and social network information and online network activity to determine and develop resiliency metrics. Impacts of the Solution to the Grand Challenge Many participants stated that a critical need for evaluating community resilience is an understanding of “baseline” measures of resilience to enable measurements of the changes in resilience that take place over time. Such changes could include reductions in expected losses that accompany the adoption and implementation of new codes, retrofit programs, and other improvements in the earthquake resistance of the built environment. They also could include changes in social vulnerability and resilience (such as those related to fluctuations in income levels, migration patterns, and the size of at-risk populations) and changes in exposure to risk (e.g., due to decisions to develop or to restrict development in hazardous areas). Participants stressed that observatory networks are needed, in part, because vulnerability and resilience are continually in flux. Research and methodological approaches that take into account ever-changing community vulnerability and resilience profiles, in their view, would be genuinely transformational. The lack of longitudinal resilience data makes it difficult or impossible to determine whether measures that are intended to reduce future losses actually make a difference. Such data could also reveal social factors that affect resilience independent of the kinds of engineering advances that were emphasized during the workshop. Participants noted that the impact of a more holistic framework for measuring, monitoring, and evaluating community resilience could be enormous. Better models and data for understanding community resilience could facilitate more effective decision making, and in turn, improved community resilience. Validated profiles of community performance that result from detailed and rich datasets from past events could enhance the confidence of decision makers in the tools and methodologies developed by the research community. This, in turn, could enhance their use both before and after major earthquakes.
DECISION MAKING Description of the Problem Many workshop participants noted that achieving earthquake resilience requires a community-based, holistic approach that includes decisions and actions that are based on overarching goals, a clear understanding of the built environment, rapid and informed assessment data, and planned reconstruction and recovery processes. Communities typically build based on traditional standards, and when affected by major earthquakes, respond and recover based on intuition, improvisation, and adaptive behaviors that are drawn from the individuals available to participate. Consequently, the lessons learned in one community and event rarely translate to the next community affected. In order to facilitate better decision making, participants explained, meaningful data are needed that allow end users to quantify current and improved levels of community resilience. They stressed the importance of using historical data when testing and validating strategies for translating the results of quantitative and qualitative studies on community resilience. Characteristics of the Grand Challenge An observation reiterated during the workshop is that research on community resilience has not made a significant impact on the decisions and actions of decision makers. Prior to a seismic event, for example, interest in seismic mitigation and preparedness is often limited or non-existent. Immediately following an event, the environment within which decisions must be made by first responders and the public can be chaotic and complex, hindering optimal decision making. A number of participants suggested that the scientific and engineering community should explore the complexities of these operational environments and how they evolve on multiple length and time scales. They also expressed a need for basic research to explore a variety of mechanisms for motivating action, including: • Providing information and developing incentives for action. • Developing simulation-based decision-making strategies for use in community development, pre-event planning, early post-event response, and through the long-term recovery process. • Providing incentives to promote community development and pre-event planning. • Using state-of-the-art decision-making tools that would lead to more efficient resource allocations. • Developing methodologies and tools that allow decision makers to compare different strategies for post-earthquake reconstruction and long-term preearthquake mitigation.
15
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
• Performing case studies that demonstrate the efficacy of proper planning and response after damaging events. Participants suggested that exploring how society approaches preparedness and post-event response and recovery could help ensure that lessons are learned from past seismic events and applied to community development and rebuilding, requiring a transition to a community-based risk management and resilience paradigm. Transformative Approaches to the Solution Many participants expressed a need for more fundamental research on decision making under conditions of uncertainty, and decision making for low-probability/highconsequence events, along with basic research and research integration in areas such as public administration and public policy, communication theory and practice, knowledge and technology transfer, and decision science. They suggested that this research relate both to pre-event (planning, construction standards, prioritization, simulators, training) and post-event scenarios (emergency response and recovery). Ubiquitous sensor data would be required to drive the decision support engines. In the post-event period, heterogeneous inputs and outputs from a range of linked simulation systems (coupled with field sensor data) could be managed and assessment information used to inform first responders for efficient resource allocation. Precise quantitative assessment of the damage state would be critical, along with an assessment of the impact of damaged systems/components on other interdependent systems. A partnership between engineers, scientists, and the emergency management community could remove barriers for adopting new technologies, and cyberinfrastructure could support the near-real-time delivery of information that supports post-event recovery activities. New information technologies could allow decision makers to share information across organizations without the threat of security leaks or breaches. Many participants noted that efficient and accelerated recovery requires timely post-event repair and rebuilding decisions that take into account models and tools to forecast long-term consequences and the impacts of potential mitigation options (such as land buyouts, redesign/reconstruction changes). In their view, current models do not adequately reflect longer-term cascading impacts of large-scale disasters, and the resilience of repair technologies is not well understood. An effective system of post-disaster mitigation and recovery assistance could utilize a “resilience basis” to determine best use of public funds. Participants suggested that the integration of research on risk communication and decision making with methods developed for resilience assessment, including simulation and visualization studies, could lead to new approaches to planning and stimulate action. Deployment could involve
expanded technology transfer that includes education Â�studies to facilitate close collaboration between researchers and decision makers. Application of advanced information technology (e.g., cloud computing, apps, and HTML5-enabled web) and social networking-style approaches could help to improve resiliency communication, education, and decision support. Participants also noted that the current links between ubiquitous data streams, high-fidelity modeling, and effective decision making are weak or non-existent. These connections could be enhanced by the development of an integrated system that identifies events, creates and monitors real-time data, updates models, incorporates crowd sourcing technologies, and informs decision makers. Real-time assessment of damage to buildings and infrastructure could help in defining effective recovery strategies that emphasize the rebuilding of community sectors that promote rapid economic as well as social development. This could lead to a paradigm shift away from solely engineering solutions to a holistic suite of resilience options including land use planning, performancebased construction standards, and different configurations of post-event reconstruction. However, participants also noted the challenges involved with developing such a system—the linkage between technological solutions and effective decision making would need to address a number of fundamental social science and policy questions (e.g., in the context of competing community needs, when is the most appropriate time to promote an earthquake resiliency policy agenda?). Participants stressed the importance of developing an integrated system that addresses loss reduction, decision making, and complex cognitive, social, political, and economic dimensions in this process. Impacts of the Solution to the Grand Challenge Workshop participants highlighted potential impacts associated with meeting this Grand Challenge. Comprehensive support engines for decision makers would likely lead to significant savings of lives and losses, transformative potential for training and educating the next generation of professionals, direct dissemination of research into practice, more rapid and accurate post-earthquake assessments (in terms of both response and recovery), and measurable output that allows decision makers to track and evaluate the impact of their decisions.
SIMULATION Description of the Problem Participants noted that large-scale seismic events pose countless safety and logistical challenges to dense urban communities populated with both people and critical infrastructure systems. Many dense urban centers have grown over decades with multiple stakeholders involved in the planning, construction, and management of the infrastructure
16 systems that support the economic prosperity and quality of life of the society. With these considerations in mind, workshop participants discussed the need for a new generation of high-performance simulation technologies to accurately forecast the physical, social, and economic impacts of largescale seismic events on dense urban regions. Characteristics of the Grand Challenge Because of the complex growth patterns in urban regions, knowledge of the inventory of infrastructure components and points of connection between different infrastructure types is lacking (NRC, 2011). Although individual state agencies and utilities may maintain databases of their infrastructure systems, many of these databases are proprietary and use a myriad of database standards, making interÂ� operability challenging. Provided that infrastructure systems are interconnected and have vulnerabilities at their points of interÂ�connection, there is an opportunity for new methods focused on autonomously creating a comprehensive database that provides a complete mapping of all infrastructure systems in a region. Hence, a number of workshop participants stressed the need for scalable tools that Â�autonomously Â�create an accurate database of all infrastructure components, including points of interdependency with other infrastructure components. Empowered with this complete mapping of an urban center’s infrastructure systems, powerful simulation technologies could model the time and spatial impacts of a seismic event at all length scales, spanning from the component scale to the regional scale, and from disaster response to community recovery. Transformative Approaches to the Solution Several participants noted that to effectively address the Grand Challenge, wide gaps of scientific and engineering knowledge will need to be bridged to create transformative solutions. These gaps were highlighted in the presentation by Omar Ghattas2 and in the “pre-event prediction and planning” breakout group discussions. For example, it will be important to explore new technologies aimed toward the creation of a comprehensive urban infrastructure database from both proprietary data sources (e.g., utilities’ inventory databases) as well as from analysis of socioeconomic data sources (e.g., census data, economic indicators). The end result will be the creation of an infrastructure “genome,” much like the genome used to map the fundamental protein structures that make up life. Tools that allow the genome to evolve with the growth patterns of the urban region itself could be created to ensure long-term accuracy and validity. Powerful new forms of multi-scale computing architectures 2
See the white paper in Appendix B by Omar Ghattas, the keynote Â� speaker on Modeling and Simulation: “Uncertainty Quantification and Exascale Computing: Opportunities and Challenges for Earthquake Â�Engineering.”
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
could be created to link a heterogeneous array of simulation tools to provide a complete toolset for regional simulation of the impact of an earthquake and tsunami. High-performance computing technology that enables repeated simulation for stochastic modeling of earthquake responses and community responses would likely be a key technology. To update and verify the multi-scale simulation environment, the data generated by sensors embedded in the built environment for both seismic and infrastructure monitoring may be explicitly utilized. A sensor fusion approach could incorporate other forms of data including data derived from remote sensing technologies and crowd sourcing datasets. Impacts of the Solution to the Grand Challenge Participants discussed the enormous potential benefits of such a rich and expressive simulator toolbox. A means of autonomously creating an accurate inventory of infrastructure systems without relying upon the sharing of information from the many owners of the infrastructure components, some noted, would offer the engineering and social science community an unprecedented opportunity to utilize inventories of infrastructure systems that enable regional modeling of the short- and long-term impacts of large earthquakes and tsunamis. The tools that link simulation across multiple length and timescales could enable predictive modeling that could shape the community’s efforts in preparedness yet Â�allow emergency response officials to create optimal plans that most efficiently allocate their scarce resources immediately after an event. Furthermore, simulation of how infrastructure systems are interdependent, both in operation and failure, could provide a wealth of new knowledge on how complex, regionally distributed infrastructure systems are vulnerable to regionally destructive events such as Â�tsunamis. Beyond earthquake engineering, fundamental science aimed toward linking heterogeneous simulation tools that incorporate physical models with the simulation of community response to disasters could facilitate discovery for other forms of natural and man-made hazards.
MITIGATION Description of the Problem Community resilience, as described by participants in the “pre-event prediction and planning” breakout group, fundamentally depends on developing risk assessment and mitigation strategies for the renewal and retrofit of the infrastructure sectors most highly vulnerable to earthquakes and tsunamis. These sectors include water and wastewater supply and distribution systems, power and energy infrastructure, communication systems, transportation systems, at-risk buildings, and coastal communities in seismic zones. A number of participants noted that improvement in mitigation requires proactive changes in public policy that facilitate new
17
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
strategies for safe and robust design and construction; proposals for innovative funding strategies to upgrade vulnerable sectors of the built environment; and a range of options available to create resilient designs for both existing and new systems. Consequently, it would be important to document the current vulnerabilities within the built environment; to prioritize the most crucial mitigation needs; and to develop cost-effective and sustainable mitigation strategies that are embraced by the communities at risk.
• • • •
Characteristics of the Grand Challenge A large earthquake or tsunami in a highly populated region of the United States would cause massive damage to the built environment and communities in the region. The resulting social and economic consequences would cascade across the country, particularly if major energy, transportation or supply hubs are affected. As an example, Kobe, Japan, has yet to recover completely from the 1995 Great Hanshin earthquake. Participants noted that the potential conÂ�sequences of inadequate mitigation of the built environment’s most vulnerable sectors are acute. Therefore, this Grand Challenge includes developing strategies for identifying and prioritizing the sectors of the national built environment that are most vulnerable to catastrophic losses from earthquakes and Â�tsunamis, in addition to developing approaches for renewal and retrofit of these sectors to ensure a safer environment and a more resilient community. As important, meeting this Grand Challenge involves providing fundamental strategies that mesh with related national priorities—such as ensuring national competitiveness and economic growth in key Â�regions of the country that are vulnerable to seismic risk—as well as enabling new solutions for reviving the built infrastructure to ensure more sustainable and secure communities. Key characteristics of this Grand Challenge include developing strategies to measure, monitor, and model a community’s vulnerability, motivations, and mitigation strategies, and establishing mitigation solutions for its most vulnerable sectors. Strategies could be based on the use of a new generation of simulation tools and design solutions coupled with up-to-date information available from distributed sensing systems. Individual participants noted that these strategies would require: • Accessing an accurate inventory of built assets (e.g., buildings, lifeline networks, socioeconomic data, policy data, natural environment, and topology). • Understanding the scope of possible seismic and tsunami hazards, including the range of likely magnitudes, locations, recurrence intervals, and ground motion characteristics. • Developing advanced simulation tools that provide a range of information on potential catastrophic consequences of scenario seismic and tsunami events
• • • •
based on forecasting of future inventory, population dynamics, and trends in design and construction. Modeling interconnected and interdependent distributed systems, including lifelines. Accessing distributed sensor sets to update model parameters to ensure accurate data for simulations, and coupling simulations with sensor inputs. Establishing a broad range of performance metrics to ensure decisions related to mitigation priorities. Integrating uncertainty modeling to facilitate informed decisions. Developing quantitative approaches that facilitate incorporation of individual and organizational motivations for promoting mitigation. Modeling public and private funding strategies for mitigation to enable thorough assessment of options. Developing aggregate inventories of community risk/ resiliency for use in land use planning and emergency planning. Integrating mitigation strategies and new design solutions to reduce seismic and tsunami risks that incorporate new developments in sustainable mate� rials and technologies.
Transformative Approaches to the Solution Participants suggested that the most effective strategies for assessing risk and prioritizing mitigation strategies would integrate related key elements that influence decisions on renewal and facility retrofit or replacement. These elements might include, but are not limited to: • Strategic prioritization to achieve economic growth and urban redevelopment. • Regional or local security. • Public health objectives related to clean air and water. • Energy policies and priorities. • New methods of infrastructure procurement to maximize the amount of seismic and tsunami mitigation that may be achieved within limited budgets. Vulnerability assessment and prioritization of renewal options could be achieved through regional simulations and design strategies that access data from an array of distributed databases and sensor networks and link layered simulations of seismic and tsunami events. Participants indicated this could result in documentation of direct damage and socioeconomic impacts as well as enhanced performance based on scenario mitigation solutions. New mitigation solutions could enable cost-effective retrofit and renewal options for the most vulnerable sectors of the community. Open access data architecture could enable access and use of distributed databases and sensor arrays. A number of participants noted that new strategies to understand the linkages between the physics-based phenomena that lead to infrastructure damage,
18
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
and socioeconomic and policy phenomena that result, are important for prioritizing the vulnerability of sectors of the infrastructure and regional populations. Impacts of the Solution to the Grand Challenge Solving this Grand Challenge could directly enhance community resilience. New strategies for identifying the most vulnerable sectors of the built environment could help communities to better mitigate seismic, tsunami, and related cascading hazards.
• • •
DESIGN TOOLS Description of the Problem This Grand Challenge involves developing and exploiting new materials and innovative structural concepts and integrating them within design tools to improve the performance of all types of infrastructure and to increase earthquake resilience in a sustainable manner. Participants noted a wide range of sustainable, highly resilient, new materials that offer opportunities to change the way infrastructure is designed and constructed.3 Innovative types of structural systems, such as self-centering systems with replaceable fuses, could exploit these new stronger but more brittle materials. The power of PBEE could be harnessed to achieve resilient infrastructure incorporating these innovative new materials and structural systems. Participants emphasized that fundamental research is needed to extend existing PBEE techniques for buildings to cover the full range of infrastructures, including lifelines and other critical facilities. Supporting all these developments in high-fidelity testing and modeling techniques would likely achieve a high level of confidence in performance prediction for the complete range of infrastructure types.
•
•
of potential repair, retrofit, and new construction methods. Utilization of new materials and systems to develop better, less-expensive options for repair and retrofit. Many constraints (e.g., complexity) that limit the use of PBEE in the design of new structures. Extension of PBEE to incorporate a holistic assessment of fragility including the involvement of nonstructural elements, foundations and soil-structure interaction, structure, building content, services, and the adjacent buildings. A more developed PBEE could take into account multiple hazards—such as fire, tsunami, and aftershocks—and consequence functions that consider the wider societal impact of damage, including business interruptions and downtime. Reliable fragility data for the full range of infrastructure types, including bridges, lifelines, and critical structures requiring physical testing of components and complete systems (some participants stated that although such systems are complex multi-scale problems, a move away from empirical data is needed). Fundamental research to understand the influence of aging and degradation of infrastructure in order to develop appropriate fragility data for existing infrastructure.
Transformative Approaches to the Solution Participants acknowledged that achieving a high level of confidence in performance prediction for materials, subsystems, and complete structures requires the availability of high-fidelity testing and simulation techniques, encompassing:
• New validated physics/mechanics-based models for the many new materials becoming available to the earthquake engineering community. • Methodologies to assess the environmental (e.g., carbon footprint) and performance-related impact
• The development of detailed mechanics-based Â�models for modeling materials and subsystems using high-Â�performance computing or parallel computing facilities to study system behavior over a wide range of scales. • Creation of reference datasets from experimental tests for analysis comparisons and blind prediction studies to increase confidence in the numerical simulations. • Development of methods for automated validation of proposed analytical models against existing empirical datasets. • Development of software platforms and hardwarein-the-loop techniques for testing materials and structural components with realistic boundary conditions and permitting physics-based modeling of interdependencies among lifeline systems.
See the white paper in Appendix B by John Halloran, the keynote speaker on Materials: “A Built Environment from Fossil Carbon.”
A number of participants also emphasized the need for development of new and emerging materials normally
Characteristics of the Grand Challenge Participants noted that this Grand Challenge would require fundamental research into new materials and structural systems that have the potential to transform the construction, repair, and seismic performance of infrastructure. Extensive testing and modeling would be needed before such developments could be implemented within existing PBEE methodologies. Some specific challenges to be addressed include:
3
19
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
used outside the construction sector (such as ultra-high-� performance concrete, carbon products, green �binders, recycled materials, autoadaptive self-healing �materials) that could be used for retrofit and construction of sustainable yet highly resilient infrastructure systems. This would require research into innovative ways to incorporate such materials in structures, energy capture, and brittle fuses, along with research into methods for using these new materials and techniques to create economical retrofitting systems and protection systems. Participants noted that new methods would also be needed for incorporating reparability into new designs and the development of performance metrics to quantify resilience and sustainability in a holistic manner. Benchmarking could ensure reliable development of PBEE, along with better analysis techniques and statistical methods for characterization of uncertainties. The development of reliable fragility curves for bridges, lifelines, and critical systems would also be important. Extension of the building information management systems developed for building infrastructure for the modeling lifeline systems could open up new ways to characterize such systems.
Impacts of the Solution to the Grand Challenge Workshop participants noted that the development of new materials, systems, and design tools offers many opporÂ� tunities to create more resilient and sustainable societies. The composites—utilizing carbon or other materials—currently being developed and used within other engineering sectors have the potential to transform the way infrastructure is designed and dramatically improve the resilience of infrastructure systems in an earthquake. Participants expressed the view that most existing composites are not appropriate for infrastructure use, but more economical construction grade variants of these materials would still be significantly Â�stronger and lighter than standard construction materials while being appreciably more sustainable. New materials also offer opportunities to design economical retrofitting systems that could be appropriate for any community. Full acceptance and implementation of PBEE has the potential to transform the way all types of infrastructure are designed. New high-fidelity testing techniques could reduce the dependence of larger scale simulations on empirical evidence and support the development of more accurate decision support tools. The potential exists to design and build vastly improved protective systems, which might also incorporate innovative features such as energy capture from the earthquake motion.
3 Networks of Facilities
The grand challenge problems described in Chapter 2 emerged during breakout sessions and were discussed in more depth and refined during plenary sessions. In subsequent discussions, workshop participants were asked to identify general requirements for the experimental infrastructure capabilities and cyberinfrastructure tools associated with addressing the grand challenge problems. The committee consolidated the results of these discussions into descriptions of 14 distinct networks of facilities, which are presented in this chapter. A key part of the discussions involved the characteristics of the network of facilities, both experimental and cyberinfrastructure. Networking allows collaboration of geographically distributed researchers and team members as they utilize multiple facilities. Collaboration tools include real-time and asynchronous communication, access to databases, simulations, and experiments. Advanced collaboration tools and social media could allow new types of interaction to develop and enhance the educational and outreach functions of the network. Workshop participants reiterated that data storage, search, and mining are critical tools for the network. Access to simulation and analysis software from petascale computers to mobile apps could leverage a substantial toolset in earthquake engineering and other applications, and unleash developers to create new applications to meet the demands of the Grand Challenges. Real-time communication with Quality-of-Service guarantees could allow advances in hybrid simulation and advanced testing methods. Participants envisioned that the user communities for the facilities would encompass a wide range of Â�researchers, practitioners, planners, and other officials, and that the data, models, and information sources would be available and documented for the general community. In discussing the networks of facilities, participants described the characteristics of a unique community resilience observatory and an “instrumented city” testbed that would create urban-scale laboratories without walls to integrate experiments, simulations, and models of com-
munity behavior to advance knowledge about resiliency. An advanced observatory and mobile monitoring system could provide data at urban scales to researchers before and after earthquake events. Specifically, the community resilience observatory could offer researchers an opportunity to develop, test, and evaluate different methodologies for quantifying community resilience in different parts of the country; to monitor and track recovery in areas that have experienced major catastrophes; and to ensure that benchmark data for measuring resilience can be standardized across the country. In turn, products from the community resilience observatory could benefit land use planners, emergency responders, and state, regional, and local policy makers in their efforts to better prepare for earthquakes. An instrumented city, in addition, could allow researchers to integrate the output from many different sensors (from strain gauges to satellites) to monitor the performance of a city during an actual disaster. The continued collection of data—both before and after an earthquake—could allow not only researchers but also community policy makers to generate critical benchmark datasets for use in quantifying the impact of risk reduction measures for a community. The experimental facilities suggested by participants encompass testing and monitoring over a wide range of scales, loading regimes, boundary conditions, and rates on laboratory and field (in situ) specimens that would be needed to address the grand challenge problems identified during the workshop. At the material scale, facilities can generate data about the properties and behavior of sustainable material. At the full scale, facilities can provide urgently needed information about the performance of complete structures, including the effects of soil and non-structural components. The interlinking of multiple sites through methods such as hybrid simulation would allow experiments of the “whole” to be greater than experiments on the “parts.” Participants suggested that cyberinfrastructure tools are essential for capturing, analyzing, and visualizing experiments and for supporting the advanced simulations identified in the Grand 21
22
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
TABLE 3.1╇ Linkages Between Facilities and the Five Overarching Grand Challenges. Community Resilience Framework
Decision Making
Simulation
Mitigation
Community Resilience Observatory
√
√
√
√
Instrumented City
√
√
√
√
Design Tools
Earthquake Engineering Simulation Center
√
√
√
√
√
Earthquake Engineering Data Synthesis Center
√
√
√
√
√
Earth Observation
√
√
√
√
Rapid Monitoring Facility
√
√
√
Sustainable Materials Facility
√
Networked Geotechnical Centrifuges
√
SSI Shaking Table
√
√ √
√
√
√
√
√
√
√
√
Large-Scale Shaking Table
√
√
√
√
√
Tsunami Wave Simulator
√
√
√
√
√
Advanced Structural Subsystems Characterization Facility
√
√
√
√
Non-Structural, Multi-Axis Testing Facility
√
√
√
√
Mobile Facility for In Situ Structural Testing
√
√
√
√
Challenges. A simulation center and data synthesis center were identified as separate but interlinked facilities because of their very different services and capabilities. Table 3.1 shows how the facilities discussed in this chapter could address the five overarching Grand Challenges described in Chapter 2. As one example from the table, the rapid monitoring facility addresses problems described in the Community Resilience Framework, Decision Making, and Simulation Grand Challenges. The ordering of the facilities does not indicate prioritization.
COMMUNITY RESILIENCE OBSERVATORY The community resilience observatory, as envisioned by participants in the “community resilience” breakout group, would encompass interlinked facilities that function as a laboratory without walls. It could integrate experimental testing and simulations with a holistic understanding of communities, stakeholders, decisions, and motivations. The observatory could support basic research on interdependencies among systems, the multiple dimensions of resilience, and analytic tools for resilience measurement that take those interdependencies. It could host evolving community data and coordinating models that use that data to produce knowledge about resilience. Participants noted that comprehensive datasets from past earthquakes that quantify both the direct and indirect impacts of these events, empirical indicators (e.g., socioeconomic information on communities) that measure the resilience or sustainability of communities from past disasters, and tools and platforms (software or social networking solutions) that allow researchers to access and
use the above data in an open resource framework, would be especially important aspects of this data collection. Several participants noted that the concept of a resilience observatory is not new. In 2008, the National Science Foundation (NSF) and the U.S. Geological Survey supported a workshop that brought together leading Â�researchers from the disaster research community to explore the creation of a new NSF observatory focused on resiliency and vulnerability. Such an observatory would address obstacles by “(1) supporting development of long-term longitudinal datasets; (2) investing in the development of data collection protocols to ensure comparable measurement in multiple socio-political environmental settings and across multiple hazards; (3) building on and complementing existing data collection efforts and activities in the public and private sectors; and (4) enhancing the sharing of data throughout research and practice communities” (Peacock et al., 2008). The observatory concept discussed during this present workshop is similar to that of the 2008 workshop. Participants described this observatory as a virtual clearinghouse for a broad range of data that could be used to monitor, measure, and evaluate the resilience of a community. As discussed at the workshop, these data would be housed in different laboratories across the country and would be accessible by all researchers interested in studying community resilience. The observatory was also seen as a series of testbeds to study post-earthquake recovery in different parts of the country. By examining recovery in different regions, researchers could begin to evaluate the scalability of methodologies and models designed to measure community
23
NETWORKS OF FACILITIES
performance. Finally, the observatory could be used to link researchers from various disciplines in order to study community resilience from a holistic perspective. By taking part in a virtual network from different parts of the country, researchers could better study the physical and socioeconomic factors that affect community resilience.
(below), and it could leverage high-performance computing services available through national networks. System and application development would be an essential part of such a service, to create the core simulation services and interfaces needed to support further advances in the earthquake engineering community.
INSTRUMENTED CITY
EARTHQUAKE ENGINEERING DATA SYNTHESIS CENTER
An instrumented testbed in a high-risk, urban environment could provide invaluable data about a community’s resilience to earthquakes. New instrumentation from strain gauges to satellites could monitor and measure at multiple orders of scale. For complex lifeline systems—including transportation networks—participants emphasized a need for underground sensing and new monitoring devices that are wireless, self-locating, and self-placing. Leveraging other uses of lasers, imaging, satellites, and networks such as smart grids could contribute to collecting data in a region. As such, an instrumented testbed would allow capturing the response of complete, interconnected infrastructure systems and their interactions with urban systems. A constellation of sensors could be connected to a central data repository (e.g., the Earthquake Engineering Sensor Data Synthesis Center). As envisioned by workshop participants, this repository would require new technologies with respect to data management, communication, data fusion, data processing and dissemination, and data sharing. The instrumented city could allow unprecedented research on studying decisionmaking processes for development and calibration of comprehensive community models. It could be a specific site or region where many of the sensor systems described above are already in place or could be installed as part of other programs.
EARTHQUAKE ENGINEERING SIMULATION CENTER Massively parallel computers, fast memory access, and large storage in an earthquake engineering simulation center could enable high-performance computing computations for large-scale modeling and simulation. Such a center could bring together earthquake engineering researchers with experts in algorithm development, computational and statistical methods, and high-end computational and cloud development methodologies to enable transformative advances in simulation. Such a center could include theory-based simulation and multi-scale, multi-component modeling, as well as data-intensive, inverse, and hybrids of these paradigms. An interactive visualization capability could be networked and distributed for comparing simulations and experimental data. Participants noted that an important requirement is the capability for regional simulations including integrated visualization and interactive decision making. Such a simulation center could have 100 GB bandwidth network connectivity with the Earthquake Engineering Data Synthesis Center
An earthquake engineering data synthesis center, as envisioned by participants across multiple breakout groups, would offer the research community a large-scale database system for ingesting data sources from a variety of �sensor types including imaging, remote sensing, video, and information management systems (e.g., BIM, GIS). Such a �center could support the execution of models over that data to provide curated reference data, inferred derived information, simulations of normal and disaster scenarios and mitigation and response, and community services supporting data access and decision support. The center could assume federation and harvesting of data as a significant mechanism and focus on integrated, derived, and curated data products, and also offer advanced search and retrieval based on meta-data and action queries on data. Such a rich data source could help researchers understand the response of complete infrastructure systems in a region at multiple scales through networking with sensor galaxies and all experimental and field facilities. The center could provide well-defined abstractions that would empower users to develop tools for data analysis and monitoring to support statistical and inferential discovery.
EARTH OBSERVATION Many workshop participants expressed a need for integrated continuous and multi-sensor (e.g., aerial, satellite, unmanned aerial vehicle) observations of communities at various scales (e.g., buildings, neighborhoods, regions, and countries) for characterizing the physical attributes of communities and monitoring the effects of earthquakes (e.g., damage assessment and recovery). These earth observation systems could offer optical as well as dimensional views (3-D using radar and LiDAR [light detection and ranging] sensors) of cities that would quantify attributes of cities including location, type, and density of buildings; location of critical lifeline systems; and natural attributes that could contribute to the vulnerability of an area (e.g., low-lying coastal areas subject to tsunami effects). Many of these networks and systems are already in place, and a number of participants noted that existing resources could be leveraged to accomplish the above objectives. To develop a holistic solution for quantifying the vulnerability and resilience of large cities, many participants stressed the importance of including a remote sensing component.
24
RAPID MONITORING FACILITY Participants noted that a rapid monitoring facility could provide the earthquake engineering community with a suite of highly portable sensing and data acquisition tools that could be rapidly deployed to structures, geo-facilities, and lifelines to monitor their stability after seismic events. Included in the deployable facility could be robotic systems that would be capable of sensor placement in partially collapsed structures and in lifeline systems with tight, difficultto-reach locations. Sensor arrays deployed into critical infrastructure systems could provide a wealth of response data during aftershocks, providing valuable data for future modeling.
SUSTAINABLE MATERIALS FACILITY There is an emerging range of new, sustainable, highly resilient materials that offer opportunities to change the way infrastructure is designed and constructed. Many of these high-performance materials are being developed for the aerospace and mechanical industries, and are not currently appropriate for adoption by the construction industry (because of very high prices and limited availability). Participants noted that there is a significant opportunity to partner with material science facilities to develop and test new construction-grade materials, which might be self-healing, capable of energy capture, or ultra high strength, and to understand the use of sustainable materials for earthquake engineering applications. Although existing materials facilities might be appropriate for some of this development, it is likely that augmented or new facilities would also be needed to test these materials under the conditions they are likely to experience when used in construction, accounting for the influence of aging and degradation.
NETWORKED GEOTECHNICAL CENTRIFUGES Multiple networked geotechnical centrifuges, each including innovative capabilities for robotic manipulation and actuation within the centrifuge container during the experiment, could allow new types of experimental modeling of landslides (including submarine), liquefaction, and tsunamis. Unique hybrid simulations would be possible through networked facilities, thus enabling a more detailed assessment of interaction effects between structures and foundation systems and large-scale integrated geotechnical failures.
SSI SHAKING TABLE A large-scale, dynamic shaking table designed for soilstructure interaction (SSI) experiments, as envisioned by participants in the “design of infrastructure” group, would enable a significant throughput of SSI experiments to help
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
advance knowledge of this crucial component of earthquake engineering. A large-scale testing system could facilitate studying of the interaction of geotechnical conditions for both infrastructure components as well as building systems. Self-organizing wireless sensors, as well as other new types of sensing strategies specific to SSI, could enable high�resolution assessment of progression of damage in SSI systems and the development of new strategies for more �robust design of structures and infrastructure systems. Hybrid simulation could also provide the realistic, time-dependent loading on specimens that is important for accurate assessment of soil-structure interaction.
LARGE-SCALE SHAKING TABLE A large-scale shake table facility capable of full-scale structural testing was viewed by a number of workshop participants as being important for addressing the Grand Challenges. They noted that there are significant knowledge gaps about structures that are damaged or partially collapsed and the modes of failure. Testing complete structures or full-scale subsystems in multiple directions would allow improved understanding of the response of actual construction and the contributions of lateral and gravity load-resisting systems and non-structural systems. Such a facility could provide fundamental knowledge for understanding the complete system behavior, validating post-earthquake evaluation methods for damaged structures. This knowledge in turn could help determine which structures are safe to occupy and which ones need to be demolished. As envisioned, this facility would require multifaceted testing capabilities, including hybrid methods, with the capacity to test to collapse. Workshop participants discussed the need for a study about whether it is most effective to construct a new full-scale shaking table or develop international partnerships, such as a partnership with E-Defense in Japan.
TSUNAMI WAVE SIMULATOR The tsunami wave simulator described by several workshop participants would be a revolutionary new facility that combines a tsunami wave basin with the capability to shake the ground to simulate liquefaction and subsidence. Participants noted that fundamental knowledge about largescale coupling between soil-structure and fluid interaction is lacking, and a combined tsunami and liquefaction wave tank could provide researchers with a better understanding of foundation weakening, scouring, and structural failure, which in turn would lead to improved protection for coastal communities. The wave simulator basin would be on the order of at least 150 feet wide by 250 feet long, with enhanced absorption boundary conditions capable of tsunami generation, propagation, and reproduction of local effects on coastal structures.
25
NETWORKS OF FACILITIES
ADVANCED STRUCTURAL SUBSYSTEMS CHARACTERIZATION FACILITY Many participants noted that to enable the development of more accurate structural models, a networked set of equipment that replicates the effects of corrosion, accelerated aging, and fatigue is needed for the characterization of subsystems, components, and materials. Such a facility could have the capability for multi-axial loading, high-temperature testing, and high pressures. It would need to be able to test full-sized or close-to-full-scale subsystems and components under fully realistic boundary and loading conditions, including rate effects to avoid issues with scaling, and would need to be supported by a comprehensive set of high-performance instrumentation. Such a facility could enable the development of high-fidelity physics-based models for incorporation into simulations of complete structures. It could also enable the characterization of the full lifetime performance and sustainability of structural elements and materials and allow the development of appropriate retrofit and strengthening techniques for existing aging infrastructure.
NON-STRUCTURAL, MULTI-AXIS TESTING FACILITY A significant proportion of the losses following an earthquake are the result of indirect damage to the contents of buildings, rather than damage to the structural frame. A number of participants noted that the requirements of the current seismic qualification codes cannot be fully met with existing facilities,1 highlighting the need for a highperformance multi-axis facility with the frequency range and levels of motion necessary to investigate and characterize the performance of non-structural elements (e.g., partitions) and other content (e.g., shelving, IT equipment, lighting, electrical and mechanical equipment) within a building or other
1 For example, IEEE Standard 693-1997, which contains recommended practices for seismic design of substations, cannot be met without significant filtering of the low-frequency content of the signal (Takhirov et al., 2005).
infrastructure. Such a facility would need to deliver very high displacements, velocities, and accelerations so that it could simulate the behavior of floors at any point within a building; however, it may not need to have a very high payload capacity because most non-structural items within buildings are relatively light. Such a facility could permit the development of complete building models, including the building content, and also the development of more robust non-structural elements and equipment that would be significantly less likely to be damaged in an earthquake.
MOBILE FACILITY FOR IN SITU STRUCTURAL TESTING A mobile facility for in situ structural testing, as described by participants in the “design of infrastructure” group, could be equipped with a suite of highly portable testing equipment including shakers, actuators, sensors, and high-resolution data acquisition systems that could be used to test structures, lifelines, or geotechnical systems in place. Examples include modal shakers to introduce dynamic loads on structures, bridges, and soil systems. Additional capability could include large-capacity broadband dynamic seismic wave sources coupled with improved sensing capabilities to allow the high-resolution subsurface characterization essential for regional modeling. Hydraulic actuators capable of in situ lateral loading could provide an experimental capability of testing structures. Intentional and repeatable dynamic loading of buildings, bridges, and other structural systems could allow systems to be dynamically characterized for improved modeling capabilities. Dynamic excitation of geotechnical systems could improve understanding and the modeling of liquefiable soils.
References
Peacock, W.G., H. Kunreuther, W.H. Hooke, S.L. Cutter, S.E. Chang, and P.R. Berke, 2008. Toward a Resiliency and Vulnerability Observatory Network: RAVON. HRRC report 08-02-R. College Station, TX: Hazard Reduction and Recovery Center, Texas A&M University. Takhirov, S., G. Fenves, E. Fujisake, and D. Clyde, 2005. Ground Motions for Earthquake Simulator Qualification of Electrical Equipment. PEER 2004/07. University of California at Berkeley: Pacific Earthquake Engineering Research Center.
NAE (National Academy of Engineering), 2008. Grand Challenges for Engineering. Washington, DC: National Academies Press. NEHRP (National Earthquake Hazards Reduction Program), 2008. Strategic Plan for the National Earthquake Hazards Reduction Program: Fiscal Years 2009-2013. Available at nehrp.gov/pdf/strategic_plan_2008.pdf. NRC (National Research Council), 2011. National Earthquake Resilience— Research, Implementation, and Outreach. Washington, DC: National Academies Press.
27
Appendix A Breakout Session Presentations
The summary outputs from breakout sessions that are listed in this Appendix were prepared by breakout group participants and they do not represent conclusions or recommendations of the committee or the NRC. The grand challenge problems �described here resulted from discussions in breakout group sessions, but they do not reflect the consensus views of the workshop session breakout groups.
COMMUNITY RESILIENCE (CR) Moderator: Kathleen Tierney Rapporteurs: Ron Eguchi, Laurie Johnson Participants: Mehmet Celebi, Jon Heintz, Laurie Johnson, Kincho Law, Peter May, and James Myers Grand Challenge Problem: Framework for Measuring, Monitoring, and Evaluating Community Resilience (1CR) • Description of Problem o To ensure that past and future advances in building, lifelines, urban design, technology, and socioeconomic research foster community resilience at multiple scales. • Characteristics of Grand Challenge o Resilience is multi-dimensional, multi-scale, multi-disciplinary, and integrative. It involves complex interactions that are difficult to measure, particularly in light of the multi-scale nature of resilience-related phenomena. o We are unable, at this time, to envision and measure multiple dimensions of resilience. What is needed is a characterization of resiliency in terms of scale and metrics that are applicable both for diverse systems and for their interdependencies. o It is also difficult to determine when resiliency has been achieved. Current engineering approaches do not do an adequate job, either of characterizing resilience outcomes or characterizing those outcomes in ways that are meaningful for end users. o There is a need to leverage state-of-the-art concepts and methods from multiple disciplines, including economics (computable general equilibrium models), sociology, and community psychology (social capital measures). • Transformative Approaches to Solution o Undertake a campaign of basic research on describing, defining, and quantifying/measuring community resilience. Different communities will vary in terms of “resilience profile” or “resilience portfolios,” making it possible to identify gaps that require specific interventions. o Create a resiliency observatory system (e.g., RAVON—Resilience and Vulnerability Observation Network), similar to other networks (NEON, WATERS). The network will enable data collection, integration, and monitoring across the United States, providing pre-event and post-event composite and multi-dimensional indicators of resilience. o Secure funding from a consortium of agencies, including the National Science Foundation (NSF), U.S. Geological Survey (USGS), Federal Emergency Management Agency (FEMA), and Department of Homeland Security (DHS) to support large-scale, long-term research. What policies and modifications should be put in place to help communities become more resilient? These might provide location-specific research and a unique portfolio of solutions developed for each place/problem. 29
30
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
o Establish research coordination networks (RCNs) that will be responsible for integrating research across domains and dimensions of community resilience, encompassing natural, built environment, and socioeconomic systems. • Facilities o There must be both distributed and networked facilities and “observatories.” o The design must take on a concept of a “laboratory without walls.” o Need a capability to integrate experimental testing and simulations with a more holistic understanding of communities, stakeholders, and decisions/motivations. o Need basic research on interdependencies among systems and dimensions of resilience, and analytic tools for resilience measurement that take those interdependencies into account. o Need a “hosting” capability that can accommodate evolving community data and coordinating models that use that data to produce further-derived data. o Cyberinfrastructure for laboratories without walls. Develop information technology to support the concept of a laboratory without walls, linking field tests/observations with experimental and simulation studies. • Impact of Solution to Grand Challenge o Will provide first-ever holistic approach to resilience measurement at community scale, as well as guidance for specific interventions to enhance resilience. o User expectations will be clarified, or even improved/enhanced, and will improve practice/applications. • Projects Addressing the Grand Challenge o Define best practices in resilience methods. Inventory the resilience methods/studies/metrics performed at different scales (communities, regions) and for different community components (buildings, lifelines, social networks, economy). o Define structural resilience. Use performance-based engineering (PBE) technologies to develop building performance objectives that can be associated with resiliency objectives. o Define lifeline resilience. Use PBE technologies to develop lifeline performance objectives that can be associated with resiliency objectives. This could be a project for every lifeline system. o Develop and test resiliency metrics. Research, develop, and test various methods for quantifying resilience and determine the best method for stakeholder decision making. o Investigate infrastructure interaction effects. Perform basic research including full-scale testing and simulation of buildings and lifeline systems to investigate interactions. o Create a resilience observation pilot study. It might be a candidate city, neighborhood, or group of buildings (Sample City Concept). Set a baseline and observe actions/changes over time to define metrics and timeframes of resiliency dynamics. o Development of a city as a resilient city (e.g., Hayward, Seattle, or Los Angeles). Include: building inventory, Â�instrumentation for ground motions and structures, vulnerability and loss estimation studies, and yet-to-be-developed methodologies. Assessments of upgrade requirements off the built environment for a resilient city. State of the organization and state and local governments, and disaster-related organizations. Documentation as part of the observatory. o Multi-scale simulation modeling. Develop simulation models that link the performance of buildings and lifelines to communities. o Develop data-intensive methods for using public and social network information and online network activity of all sorts to determine and develop resiliency metrics. o Quantitative recovery modeling. Develop theoretically and empirically based models of post-earthquake recovery processes. Models should be integrated across dimensions of recovery (infrastructures, housing, business/commercial facilities, public institutions, social/economic processes); should be incorporated into simulation models that forecast recovery rates and patterns after major earthquakes; and should consider resilience, adaptation, and sustainability. o Develop the base model of a city by using remote sensing of existing infrastructure, inventory, and condition. Network model interdependencies and identify regions subject to cascading engineering failures. Grand Challenge Problem: Motivating Action to Enhance Community Resilience (2CR) • Description of Problem o Research has yielded numerous findings related to community resilience, yet many of these findings are not influencing decisions and actions on the part of key decision makers such as private-sector facility owners and public-sector institutions. There is a need for basic research to explore a variety of mechanisms for motivating action, including (but not limited to) providing information and developing incentives for action that are meaningful to various constituencies, ranging from laws and regulations to informally applied norms.
APPENDIX A
31
• Characteristics of Grand Challenge o Requires advances in fundamental research on decision making under conditions of uncertainty and decision making for low-probability/high-consequence events. o Addressing this Grand Challenge also requires basic research and research integration in areas such as public administration and public policy, communication theory and practice, knowledge and technology transfer, and decision science—with engineering. • Transformative Approaches to Solution o Integration of research on risk communication and decision making with methods developed for resilience assessment, including simulation and visualization studies. o Expand technology transfer to include education. Include emergency response team training in high school. o Studies involving collaboration between researchers and decision makers. o Application of web2.0 and social networking-style solutions to resiliency communication /education/decision support (i.e., Earthquake Zillow). • Facilities o Laboratories that enable decision-focused research, e.g., experiments, simulations. • Impact of Solution to Grand Challenge o User expectations will be clarified, or even improved/enhanced, and will improve practice/applications. • Projects Addressing the Grand Challenge o Long-term post-earthquake recovery studies. Long-term post-earthquake evaluations of community recovery and reconstruction; identify key indicators of recovery. o Data fusion for decision models. Develop data fusion methods for integrating multi-dimensional, multi-scale, multimedia experimental simulation and observation data to support community decision-making processes. o Conduct historical and comparative research on the role of various “boundary organizations” (BOs) as change agents, identifying factors that contribute to the effectiveness of such organizations. BOs serve as a knowledge transfer organization. Examples include: Applied Technology Council (ATC), National Institute of Building Sciences/Multihazard Mitigation Council (NIBS/MMC), Cascadia Region Earthquake Workgroup (CREW), and California Seismic Safety Commission.
PRE-EVENT PREDICTION AND PLANNING (PR) Moderator: John Egan Rapporteur: Jerome Hajjar Participants: Raymond Daddazio, Gregory Deierlein, Steve French, Omar Ghattas, Muneo Hori, Jonathan Stewart, and Solomon Yim Grand Challenge Problem: Develop a National Built Environment Inventory (3PR) • Description of Problem o Develop an accurate, distributed, comprehensive national built environment inventory and socioeconomic database to enable dynamic forecasting of existing and future inventory • Characteristics of Grand Challenge o Inventory components ® Properties of the earth (as part of our inventory!) ® Building fabric ® Infrastructure fabric (security access is often challenging) ® Community/human functions ® Natural environment o Dynamic forecasting ® Ground truthing is not possible across large scales ® Model damage in future scenarios (i.e., based on future inventories) o Data obsolescence and heterogeneous data formats o Public policy related to data access and security ® Secure public information: utilities, military installations ® Private information (corporate, personal)
32
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
• Transformative Approaches to Solution o Development of new strategies for dynamic forecasting inventory for the future o Development of procedures to integrate remote sensing, automated 3D photo captures; integration of BIM; automated capture of transactional information (including subterranean) to update inventory o Establish access to all requisite information • Impact of Solution to Grand Challenge o Fundamental to achieving accurate high-fidelity simulations (regional, component) o Facilitates decision making based upon accurate data o Provides a strong link between engineering solutions and community and individual needs and outcomes • Projects Addressing the Grand Challenge o Develop a national inventory of buildings, infrastructure, environment, and community with appropriate attributes based on remote sensing, government records, etc., that is routinely updated by the transactional processes of the owners of the data sets o Develop ways to forecast future building and infrastructure inventories for metropolitan areas based on existing population and employment forecasts o Study jurisdictional challenges and public policy constraints for accessing data Grand Challenge Problem: Multi-Scale Seismic Simulation of the Built Environment (4PR) • Description of Problem o Enable high-fidelity simulation of the response of the national built environment to catastrophic earthquakes and related events (e.g., tsunami, fire) • Characteristics of Grand Challenge o Build models from underlying principles, physical, chemical, biological, societal o Enable holistic approach to component and system modeling; multi-hazard/hazard sequencing/cascading effects o Address interactions between multiple scales o Experiments required for validation—how to validate at scales where experiments are not possible? o Account for systematic uncertainties o Significant information technology (IT) challenges and opportunities: distributed, collaborative, confederated, stochastic o Outcome should incorporate community, social, economic outcomes o Harness simulation scenarios of other disciplines (e.g., climate change, weather modeling, etc.) • Transformative Approaches to Solution o Multi-disciplinary strategies are vital: from nano to global, physicists to public policy/social scientists o Simulations driven by economic growth modeling, not just hazard mitigation or loss or recovery modeling • Impact of Solution to Grand Challenge o Provides new knowledge on complex system interactions o Facilitates understandings at scales at which conducting experiments is unlikely • Projects Addressing the Grand Challenge o Assess and quantify first principle models for all identified critical simulations; verify against finer scale simulations: rupture-to-rafters; subsidence; collapse simulation o Validate against larger scale and/or case histories, particularly for systems-level analyses, whereby the simulation drives the project, with the test validating the simulation o Develop a systems approach to link heterogeneous simulation components Grand Challenge Problem: Integrated Seismic Decision Support (5PR) • Description of Problem o Develop simulation-based, automated, decision-making strategies for use in pre-event planning, real time during an event, in early response, and through to long-term recovery • Characteristics of Grand Challenge o This can relate to both pre-event (prioritization, simulators, training) and post-event predictions (emergency response) to fuel decisions o Decisions often need to be made under critical time constraints, especially post-event, but also pre-event o Ubiquitous sensor data are required to drive the decision support engines
APPENDIX A
33
o Must manage heterogeneous inputs and outputs from a range of linked simulation systems (coupled with field sensor data) • Transformative Approaches to Solution o Linking ubiquitous data streams with high-fidelity modeling o Nonlinear optimization strategies for complex decision support o Model reduction (from supercomputers to handhelds) o Incorporate end users in tool development to ensure effective solutions • Impact of Solution to Grand Challenge o Comprehensive support engines for decision makers o Potential significant savings of [lives and] losses o Transformative potential for training and education of professionals o Direct dissemination of research into practice • Projects Addressing the Grand Challenge o Develop decision support tools across different timescales, which enables different levels of pre-event decision support and planning o Carefully designed experiments with real users to understand human interaction with this type of system, thus feeding results into the support system o Develop an integrated early warning system, based on distributed sensors, distributed datasets, distributed personnel, distributed dissemination with appropriate community response Grand Challenge Problem: Risk Assessment and Mitigation of Vulnerable Infrastructure (6PR) • Description of Problem o Develop risk assessment and mitigation strategies via retrofit and renewal of the most highly vulnerable sectors of the infrastructure (e.g., water supply and distribution systems; power systems; communication systems; hazardous existing buildings) • Characteristics of Grand Challenge o Quantify the scope, scale, and priorities of the problem: what are the key vulnerabilities to the community (e.g., redundancy) o Accurate inventory o Understand effects of ground failure o Model interconnected distributed systems o Establish broad range of performance metrics o Harness ubiquitous sensor data streams o Link to public health objectives o Alternate methods of infrastructure procurement, as costs of retrofit and renewal are prohibitive • Transformative Approaches to Solution o What’s the silver bullet here? Open-source approach to data interaction o Need matrix organization for an interdisciplinary solution o Link to other initiatives (e.g., clean water security and health) • Impact of Solution to Grand Challenge o Increase predication accuracies of design decisions (reduce contingencies) o Directly enhance community resilience • Projects Addressing the Grand Challenge o Risk assessment to assess most vulnerable buildings o Risk assessment to assess infrastructure components, including complex interactions of interconnected networks Grand Challenge Problem: Protect Coastal Communities (7PR) • Description of Problem o Protect coastal communities from tsunamis and associated coupled multi-hazard risks to increase resilience of critical structures and communities (e.g., ports and harbors; power plants) against combined ground shaking, tsunami, fire • Characteristics of Grand Challenge o Scaling issues of multi-physics problems of tsunami generation, propagation, run-up, draw-down, and fluid-structure interaction with local structures
34
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
o Understanding cause and effect of sedimentation, erosion, and scouring o Loading on ship facilities and coastal structures • Transformative Approaches to Solution o Coupled experimental and computational simulation for ground shaking and tsunami generation • Impact of Solution to Grand Challenge o Increase resilience of critical structures and communities against combined ground shaking, tsunami, fire • Projects Addressing the Grand Challenge o Basin modeling and simulation considering boundary conditions and generation source and absorption o Resilience of ports and harbors to sequential ground shaking (multiple events) and tsunami forces o Resilience of power plants located close to the coast o Computational and experimental simulation of submarine landslides and associated community impacts • General Geo/Tsunami Facility Characteristics o Centrifuges: robotic manipulation in flight; landslides; tsunami generation; submarine landslides o Shake tables: assembly line, high-throughput shake table testing o Six-degree-of-freedom large-scale hybrid testing o Dynamic testing of lifelines o Tsunami wave basin with increased width and length, enhanced absorption boundary conditions, and capability to incorporate ground shaking; need a basin on the order of at least 150’ wide by 250’ long, capable of tsunami generation, propagation, and local effects on coastal structures (including ground shaking) o Field testing: large-capacity broadband dynamic seismic wave sources coupled with improved sensing capabilities o Sensor networks and ubiquitous sensing: transportation, lifelines; autonomous, self-organizing; sensor development: lasers, imaging, satellites, wireless, self-locating, self-placing? These allow enhanced throughput that will greatly enhance effectiveness of facilities o Underground sensing of the infrastructure in heterogeneous media o Rapid mobilization for seismic monitoring of structures and geo-facilities after major events—use robotics for fast deployment o Materials science facilities • Facility Requirements for Protecting Coastal Communities o Broad description of experimental or cyberinfrastructure facility ® Tsunami wave basin with increased width and length, enhanced absorption boundary conditions, and capability to incorporate ground shaking o How will the facility contribute to Grand Challenge solutions ® Fundamental to understand large-scale coupling between soil-structure and fluid interaction (e.g., liquefaction, foundation weakening, scouring, structural failure) o Description of facility requirements ® Need a basin on the order of at least 150 feet wide by 250 feet long, capable of tsunami generation, propagation, and local effects on coastal structures (including ground shaking) o Examples of projects on which the requirements are based ® Combined ground shaking and tsunami
DESIGN OF INFRASTRUCTURE (DI) Moderator: Ken Elwood Rapporteur: Adam Crewe Participants: Ahmed Elgamal, Kent Ferre, John Halloran, Thomas Heaton, William Holmes, Kimberly Kurtis, Stephen Mahin, and James Malley Grand Challenge Problem: Regional Disaster Simulator (8DI) • Description of Problem o A comprehensive system model is needed to achieve a disaster-resilient community o Identification of problem and planning • Characteristics of Grand Challenge o Highly complex and interlinked. Data collection is challenging o Identification/inventory of infrastructure and condition—integrated with BIM
APPENDIX A
35
o Identification of cascading-type failures and how to deal with them o Integration of socioeconomic impacts • Transformative Approaches to Solution o Development of the base regional model o Smart IT, crowd sourcing o Test simulator on sample city/region o Requires input from 1CR and 3PR o Include human decision-making processes • Impact of Solution to Grand Challenge o Potential for multi-hazard simulation o Real-time decision making • Projects Addressing the Grand Challenge o Developing the base mathematical model of a region ® Developing a mathematical model of region (topology) ® What technology is needed, software, hardware ® Formulation of the problem o Development of inventory dataset (including changes with time) ® Development of technology to collect the data—remote sensing techniques ® Including current and possible future development of buildings, lifelines, socioeconomic and community resources ® Inventory and condition data o Developing models for interdependencies ® Developing appropriate network models ® Developing models for fire following ® Developing methods to identify and design for cascading engineering failures (e.g., transportation networks) potential for interlinked failures ® Models for permanent soil deformation (including for liquefaction, landslides) and its consequences ® Fault rupture modeling o Development of methods for model updating from sensor networks o Development of building models ® Occupancy, contents structure, value, proximity o Economic models o Integration of human decision models o Applications of such a simulator ® Identification of weaknesses/Making decision to mitigate weaknesses ® Pre-event drills ® Loss estimation ® Real-time response and mobilization ® What-if scenarios for mitigation o Use of example cities/regions to validate simulator models o Simulator has potential for different levels of granularity ® New generation of loss estimation tools (e.g., next generation HAZUS) o Failure of critical facilities (dam failure) Grand Challenge Problem: High-Fidelity Simulation (9DI) • Description of Problem o Achieving high confidence in performance prediction of infrastructure at scale of individual facility • Characteristics of Grand Challenge o Physics-/mechanics-based models o Holistic look at a facility (including foundations, structure, building content, services, adjacent buildings) o Multi-scale problems, complex, move away from empirical data o How to do the benchmarking? • Transformative Approaches to Solution o Reference buildings for analysis comparisons—blind prediction o High-performance/parallel computing
36
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
o Automated validation of analytical models to empirical datasets o Hardware in the loop o Integration of software platforms to permit physics-based modeling of interdependencies among lifeline systems. o Use of BIM equivalent as a transformative approach for modeling lifeline systems. • Impact of Solution to Grand Challenge o Reduced dependence on empirical evidence o Supports other grand challenges • Projects Addressing the Grand Challenge o Blind analysis—benchmarking o Physics-based modeling o Generating large-scale datasets for validating models o Instrumented buildings Grand Challenge Problem: New Sustainable Materials and Systems for Earthquake Resilience (10DI) • Description of Problem o Ensuring that sustainable solutions are also earthquake resilient. Leveraging the efforts towards sustainability for earthquake resilience • Characteristics of Grand Challenge o Assessments of environmental (carbon footprint) impact of repair/retrofit/new construction o Repair methods and how to design for repairability • Transformative Approaches to Solution o Highly resilient, new materials o Earthquake energy capture o Design of cheaper retrofitting systems that are also appropriate for global communities o The “perfect” protective system o Development of new innovative resilient buildings systems o Development of performance metrics to quantify resilience and sustainability in a holistic manner. (Redundancy is a benefit for resilience, but weakens sustainability.) • Impact of Solution to Grand Challenge o More resilient and sustainable society • Projects Addressing the Grand Challenge o Testing and modeling of emerging materials o Integrating earthquake engineering with green engineering o Evaluation of new materials ® For example, ultra high-performance concrete (UHPC), carbon products, green binders, recycled materials, autoÂ� adaptive materials, carbon footprint ® Influence of aging and degradation o The “perfect” protection system o Resilient structural concepts o Structural systems that work with brittle materials, e.g., self-centering systems, replaceable fuses o Integration with LEED o Debris management o Cost-effective retrofits o Methodology for lifecycle carbon footprint o Adaptive materials—self-healing o Energy capture materials o Structural/non-structural fuses o Design for repairability o New resilient structural concepts o Low-cost retrofits
APPENDIX A
Grand Challenge Problem: Harnessing the Power of PBEE to Achieve Resilient Communities (11DI) • Description of Problem o PBEE is a very powerful tool but has not been fully adopted o Inclusion of multi-hazards (e.g., hurricane)—No Consensus • Characteristics of Grand Challenge o Complex, multi-disciplinary, and constraints are huge against achieving o The PBEE framework exists but much more data are needed to implement it o Current process is too complicated, not practical, and not economically viable • Transformative Approaches to Solution o Significantly more benchmarking is needed for PBEE o Need for reliable fragility data including bridges, nuclear reactors • Impact of Solution to Grand Challenge o Engineering community acceptance and implementation o Transform the way infrastructure is designed • Projects Addressing the Grand Challenge o Removing roadblocks to acceptance o Better characterization of uncertainties o Developing better analysis/statistical methods o Definitions of acceptable levels of damage—all stakeholders o Fragilities including lifelines ® Experimental program o Holistic assessment of fragility ® Participation of non-structural elements, SSI, built environment o Consequence functions (links to resilient communities) ® Business interruptions, downtime o Sensitivity analysis o Optimization o Extension of PPEE to improving risk assessment o Seismic hazard improved probabilistic models o Refine risk analysis (tracking uncertainties)—No Consensus o Quantifying resilience Facility Requirements for Grand Challenges (8DI to 11DI) • High-Performance Computing (HPC) o Broad description of experimental or cyberinfrastructure facility ® Networked computers, fast, memory, storage ® Cloud computing? o How will the facility contribute to Grand Challenge solutions ® Large-scale modeling/physics-based modeling o Description of facility requirements ® Visualization tools ® AI data comparisons between simulations and experimental data o Examples of projects on which the requirements are based ® Developing the base model of a city ® Network modeling of interdependencies for regional simulator • Data Center o Broad description of experimental or cyberinfrastructure facility ® Data management, collection, visualization ® Data retrieval/documentation of data ® Data security o How will the facility contribute to Grand Challenge solutions ® Datasets for benchmark tests ® Developing relationships (materials and components) between sustainability and EQ resilience
37
38
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
o Description of facility requirements ® Cloud storage, redundancy and backups, operators and standards, metadata management o Examples of projects on which the requirements are based ® Regional model, benchmark data • Mobile Full-Scale Testing to Destruction o Broad description of experimental or cyberinfrastructure facility ® Equipment capable of destroying infrastructure (bridges, buildings, etc.) in field ® Buy buildings to test? Build a city and test it o How will the facility contribute to Grand Challenge solutions ® Generation of data for developing better whole-system models o Description of facility requirements ® Safety? ® High-performance instrumentation o Examples of projects on which the requirements are based ® Validation of high-fidelity simulations ® SSI validation • Instrumentation Repository o Broad description of experimental or cyberinfrastructure facility ® Equipment for deployment POST earthquake ® Instrumentation to pick up data from aftershocks ® Tools for damage assessment in field ® Smart pigs and drones for definition of infrastructure systems o How will the facility contribute to Grand Challenge solutions ® Collection of high-quality data from field and experimental tests o Description of facility requirements ® High-performance instrumentation o Examples of projects on which the requirements are based ® Data for real-time simulations • Advanced Subsystems Characterization Facility o Broad description of experimental or cyberinfrastructure facility ® Corrosion, accelerated aging, fatigue, multi-axial, high temperature, high pressure ® One or more machines ® Include possible interactions between ground failure and buried lifeline systems o How will the facility contribute to Grand Challenge solutions ® Allows characterization of material, subcomponent, and system performance o Description of facility requirements ® Dynamic, large scale, high load capacity ® High-performance instrumentation o Examples of projects on which the requirements are based ® Development of physics based models ® Lifetime sustainability ® Testing of components under realistic conditions • SSI Shaking Table o Broad description of experimental or cyberinfrastructure facility ® Table specifically designed for high throughput for geotechnical/SSI testing o How will the facility contribute to Grand Challenge solutions ® Producing test data for systems including SSI o Description of facility requirements ® Fast turnover of SSI tests ® Wireless sensors—self-organizing o Examples of projects on which the requirements are based • Non-Structural Testing Facility o Broad description of experimental or cyberinfrastructure facility ® Table to simulate contents of rooms, servers, etc., at various points in building ® Seismic qualification
APPENDIX A
39
o How will the facility contribute to Grand Challenge solutions ® Characterizing performance of building content ® Development of complete building models including their content o Description of facility requirements ® Very high stroke and very high velocity o Examples of projects on which the requirements are based • Hybrid Shaking Table o Broad description of experimental or cyberinfrastructure facility ® Facility to do hybrid testing o How will the facility contribute to Grand Challenge solutions ® Subsystem characterization with correct boundary conditions o Description of facility requirements ® Computing and instrumentation o Examples of projects on which the requirements are based • Summary of Network Facilities o Network of Resilience Observatories o High Performance Computing and Cloud Computing Services o Networks of dense sensor systems ® Multi-model sensor database and data center ® Distributed sensor clouds ® Citizens as sensors data fusion facility o Mobile Experimental Network ® High-capacity (shakers) for testing to destruction ® Instrumentation repository for mobile deployment o Subsystems Characterization (high capacity) Facility o Next Generation Shaking Table Network ® SSI shaking table ® Non-structural testing facility ® Hybrid shaking table ® Full-scale shaking table o Full-scale in-field testing facility
POST-EVENT RESPONSE AND RECOVERY (PRR) Moderator: Sharon Wood Rapporteur: Jerome Lynch Participants: Richard Bowker, Reginald DesRoches, Leonardo Duenas-Osorio, Mohammed Ettouney, Charles Farrar, Branko Glisic, Bret Lizundia, Sami Masri, Shamim Pakzad, Hope Seligson, and Costas Synolakis Grand Challenge Problem: Rapid Post-Earthquake Assessment (12PRR) • Description of Problem o Use assessment information to inform emergency first responders for efficient resource allocation o For a given damaged infrastructure system/component, we seek a precise quantitative assessment of the damage state o Assess impact of degraded system/component on other interdependent systems o Are there facilities that should not be entered? • Characteristics of Grand Challenges o Develop cyberinfrastructure for near-real-time data to support post-event recovery activities o Work with emergency management community to remove barriers for adopting new technologies • Transformative Approaches to Solution o Develop integrated system that identifies event, measures real-time data, updates models, and informs decision makers o Incorporation of crowd sourcing technologies
40
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
• Impact of Solution to Grand Challenge o Improve the speed and accuracy of post-earthquake assessments o Offer a more quantitative manner of assessment • Projects Addressing the Grand Challenge o Development of scalable systems to merge data from multiple and varied sources to assist first responders with search and rescue activities ® Sensing, BIM, GIS, data inundation, crowd sourcing ® Machine vision ® Smart grid o Regional damage assessment based on low-cost, dense, urban networks (consumer-based products) o Use redundant datasets in regional risk-based analysis o Paradigm shift to ubiquitous sensing (high-cost, low-density sensing to low-cost, high-density sensing) o National, wireless infrastructure to support public safety (first responders) o Cyberinfrastructure to integrate data from multiple sources (middleware) • Integration of Lifeline System Information with Ground Shaking to Identify Most Vulnerable Components o Balance statistical models with physics-based models for rapid assessment o Linking simulation and sensor output to tagging o Improve rapid assessment methods by studying long-term implications of tagging Grand Challenge Problem: Reconstruction and Recovery (13PRR) • Description of Problem o Transition to a community-based, holistic, risk management and resilience paradigm. Consider that with all the Â�disasters we have to change our approach to determining land suitability, a life space, or “urban ecosystem” in recovery • Characteristics of Grand Challenges o Lack of timeline information to make post-disaster repair and rebuilding decisions in a timely manner o Lack of models and tools to forecast out long-term consequences or the impacts of potential mitigation options (such as land buyouts, redesign/reconstruction changes) o Lack good models that reflect longer-term cascading impacts of large-scale disasters o Lack of understanding of the quality or the resilience of repair technologies o The system of post-disaster mitigation and recovery assistance lacks “resilience basis” to determine best use of public funds for achieving resilience o Lack basic understanding of user needs, workflows, and decision making post-event o Define the boundaries among acceptable, repair to pre-event condition, upgrade to higher performance level and demolition o Development of recovery simulator for affected region—economic models, material availability, and multiple Â�timescales considered in the recovery process • Transformative Approaches to Solution o Paradigm shift away from pure engineering solutions to a holistic suite of resilience options including land use planning, different uses/configurations of buildings post-event, strategic resettlement or reconstruction o Development of recovery simulator for affected region—economic models, material availability, and multiple Â�timescales considered in the recovery process o Integrate user input into program design and implementation o Assessment in real time of damage and recovery conditions of structures, infrastructure, and socioeconomic conditions, including partially collapsed structures, effects of aftershocks on damage states, and ongoing repairs o Use imaging technologies to assess damage and track key indicators of recovery over time • Impact of Solution to Grand Challenge o Develop tools to accelerate pace of community recovery and return to normalcy o Enhance understanding of community resilience by tracking recovery in a quantitative manner o Quantitative comparisons of recovery at a host of length and timescales • Projects Addressing the Grand Challenge o Set priorities for regional recovery ® Evaluation of costs at varying timescales ® Understanding factors that drive economic recovery at short-, medium-, and long-term scales
APPENDIX A
41
o Longitudinal comparisons of recovery time ® Comparisons of data from communities and decision-making processes ® Development of metrics to monitor community recovery o Experimentally verified information about seismic response of retrofitted systems ® Comparison of retrofit strategies based on cost and performance (verify FEMA 306) ® Maximize reuse of repurposed materials ® Development of new systems and strategies that are more cost-effective o Development of decision-making tools ® Merging simulation and observation, including imaging technologies, in real time to support post-event decisions in a timely manner ® Integration of sustainability tools to include economics of repair/retrofit strategies ® Enhanced speed and efficacy of post-earthquake assessment and tagging of damaged structures/systems ® Integration of sensor-based observations into the next generation of codes ® Enhanced understanding of deterioration mechanisms of infrastructure systems • Facility Requirements o Full-scale shake table ® Broad description of experimental facility — Large-scale shake table facility capable of full-scale structural testing — Capable of testing structures that are damaged or partially collapsed to observe failure ® How will the facility contribute to Grand Challenge solutions — Testing complete systems (including non-structural systems) is essential to understanding response of actual construction — What is the boundary between structures that are safe to occupy and those that need to be demolished (postearthquake assessment of buildings after a moderate earthquake)? ® Description of facility requirements — Multiple testing capabilities with the capacity to test to collapse ® Examples of projects on which the requirements are based — Verification of different types of structural systems — Post-earthquake evaluation of damaged structures o Mobile, high-capacity shaker ® Broad description of experimental facility — A high-performance shaker that is capable of being placed in decommissioned structures to apply large dynamic loads — The shaker would be mobile and can be moved from structure to structure ® How will the facility contribute to Grand Challenge solutions — Can test full-scale structures that already have some degree of degradation — Allow for non-structural components and soil to be present during dynamic testing — What is the boundary between structures that are safe to occupy and those that need to be demolished? o Instrumented city ® Broad description of experimental facility — Instrumented testbed in high-risk, urban environment ® How will the facility contribute to Grand Challenge solutions — Demonstration of new technologies with respect to data management, communication, and data fusion — Study decision-making processes for development and calibration of comprehensive, community models o Distributed sensor systems ® Broad description of experimental facility — Distributed sensor systems to capture response of complete infrastructure systems ® How will the facility contribute to Grand Challenge solutions — Actual response of complex and interconnected systems — Verification of data-based models o Multi-modal sensor database ® Broad description of cyberinfrastructure facility — Large-scale database system ingesting data sources from a variety of sensor types including traditional structural sensors as well as from non-traditional sensor streams — Ingesting data sources from inventory databases (BIM, GIS)
42
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
® How will the facility contribute to Grand Challenge solutions — Provide a rich data source to better understand the response of complete infrastructure systems ® Description of facility requirements — Data model to ensure inter-operability of data sources — Interfaces to allow for physics-based and statistics-based modeling using data — Data mining tools that support statistical discovery o Information management system ® Broad description of cyberinfrastructure facility — Information management system capable of utilizing inventory databases (BIM, GIS), ingesting data from varied sources (traditional and non-traditional), as well as real-time sensor measurements and information to inform decision makers and first responders about the condition of the community and infrastructure networks — Information available through this system can also be used to update models and provide information needed for prioritization of reconstruction and recovery efforts ® How will the facility contribute to Grand Challenge solutions — Provide a rich data source to better understand the response of complete infrastructure systems ® Description of facility requirements — Data model to ensure inter-operability of all components — Interfaces to allow for physics-based and statistics-based modeling using data — Data mining tools that support statistical discovery — Security of the data and models
Appendix B White Papers
Six workshop participants, through a keynote presentation and associated white paper, were tasked with presenting a vision that would help guide the deliberations of the workshop participants. Each discussed a key component of earthquake engineering research—community, lifelines, buildings, information technology, materials, and modeling and simulation—and considered the four cross-cutting
Â� dimensions—community resilience, pre-event prediction and planning, design of infrastructure, and post-event response and recovery. The white papers were distributed to all participants prior to the workshop, and they are published here in their original form. Final responsibility for their content rests entirely with the individual author.
43
44
TRANSFORMATIVE EARTHQUAKE ENGINEERING RESEARCH AND SOLUTIONS FOR ACHIEVING EARTHQUAKE-RESILIENT COMMUNITIES Laurie A. Johnson, PhD, AICP Principal, Laurie Johnson Consulting | Research Summary This paper is prepared for the National Science Â� oundation–sponsored, and National Research Council–led, F Community Workshop to describe the Grand Challenges in Earthquake Engineering Research, held March 14–15, 2011, in Irvine, California. It offers ideas to help foster workshop discussions on transformative earthquake engineering research and achieving earthquake resilience in communities. Over the next 50 years, America’s population will exceed 400 million, and much of it will be concentrated in the earthquake-prone, mega-regions of the Northeast, Great Lakes, Pacific Northwest, and northern and southern Â�California. To achieve an earthquake-resilient nation, as envisioned by the National Earthquake Hazards Reduction Program, earthquake professionals are challenged to strengthen the physical resilience of our communities’ buildings and infrastructure while simultaneously addressing the environmental, economic, social, and institutional resilience of these increasingly dense, complex, and interdependent urban environments. Achieving community resilience will require a whole host of new, innovative engineering solutions, as well as significant and sustained political and professional leadership and will, an array of new financial mechanisms and incentives, and concerted efforts to integrate earthquake resilience into other urban design and social movements. There is tremendous need and opportunity for networked facilities and cyberinfrastructure in support of basic and applied research on community resilience. Key ideas presented in this paper include developing better models of community resilience in order to establish a baseline and to measure resilience progress and effectiveness at an urban scale; developing more robust models of building risk/ resiliency and aggregate inventories of community risk/ resiliency for use in mitigation, land use planning, and emergency planning; enhancing efforts to upgrade the immense inventory of existing structures and lifelines to be more earthquake-resilient; developing a broader understanding of resiliency-based performance objectives for building and lifeline design and construction; building the next generation of post-disaster damage assessment tools and emergency response and recovery “dashboards” based upon sensing networks; and sustaining systematic monitoring of post-disaster response and recovery activities for extended periods of time. Envisioning Resilient Communities, Now and in the Future The National Earthquake Hazards Reduction Program (NEHRP) envisions: A nation that is earthquake-resilient
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
in public safety, economic strength, and national security. The White House National Security Strategy, released in May 2010, offers the following definition of resilience: the ability to prepare for, withstand, and rapidly recover from disruption, and adapt to changing conditions (White House, 2010). The first part of this definition encapsulates the vast majority of work that has been done under NEHRP and as part of modern earthquake engineering research and practice: strengthening the built environment to withstand earthquakes with life-safety standards and codes for new buildings and lifeline construction, developing methods and standards for retrofitting existing construction, and preparing government institutions for disaster response. The second half of this definition captures much of the recent learning and research in earthquake engineering: codes and standards that consider post-disaster performance with minimal to no disruption, as well as the linkages between building and lifeline performance and business, macro-economic, societal, and institutional recovery. But, there is much more work yet to be done, particularly in translating research into practice. What the 1994 Northridge, 1995 Kobe, and 2010 Chile earthquake and the 2005 Hurricane Katrina disasters have in common is that they all struck relatively dense, modern urban settings, and collectively illustrate varying degrees of resilience in modern societies. Resilient communities need more than physical resilience, which is best characterized by the physical condition of communities’ buildings, infrastructure, and hazard defenses. They need to have environmental, economic, social, and institutional resilience as well. They also need to do more than withstand disruption; resilient communities need to be able to rapidly recover and adapt to the new conditions created by a disaster. We are now familiar with the physical vulnerabilities of New Orleans’ levee system, but Hurricane Katrina struck a city that lacked resilience across these other dimensions as well; conditions that likely influenced New Orleans’ lack of adaptive capacity and slow recovery in the five years followÂ�ing the disaster (Public Strategies Group, 2011). Prior to Hurricane Katrina, New Orleans’ population (455,000 people in 2005) had been in decline for 40 years, resulting in 40,000 vacant lots or abandoned residences, a stagnant economy, and local budgetary challenges that severely affected the maintenance of local services, facilities, and infrastructure, most notably the school, water, and sewer systems (Â�Olshansky and Johnson, 2010). In addition, New Orleans’ social fabric was also very fragile. In 2005, the city’s median household income of $27,000 was well below the national average of $41,000, as were the homeownership and minimum literacy rates of 46 and 56 percent, respectively (compared with the national averages of 68 and 75 percent, respectively) (U.S. Census Bureau, 2000; U.S. Department of Education, 2003). The city’s poverty rate of 23.2 percent was also much higher than the national rate of 12.7 percent, and 29 percent of residents didn’t own cars (U.S. Census Bureau, 2004).
APPENDIX B
Although, in aggregate, these statistics might seem like an extreme case in community vulnerability, they are not dissimilar from some of the conditions of, at least portions of, many of our most earthquake-prone communities in southern and northern California, the Pacific Northwest, and the central and eastern United States. And, with the exception of a few pockets in northern and southern California, and Seattle, none of the most densely urbanized and vulnerable parts of our earthquake-prone communities have been impacted by a recent, large, damaging earthquake. Our modern earthquake experience, like most of our disaster experience in the United States, has largely been a suburban experience, and our engineering and preparedness efforts of the past century have not yet been fully tested by a truly catastrophic, urban earthquake. In April 2010, the U.S. Census officially marked the country’s resident population at 308,745,538, and we are expected to add another 100 million in the next 50 years (U.S. Census Bureau, 2011). This population growth is expected to be accommodated in the country’s fifth wave of migration, a wave of re-urbanism that began in the 1980s (Fishman, 2005). By the time the fifth migration is complete, it is expected that 70 percent of the country’s population will be concentrated within 10 “mega-regions” of the country (Barnett, 2007; Lang and Nelson, 2007). Half of these mega-regions are in earthquake-prone regions of the Northeast (from Washington D.C. to Boston); the Great Lakes (Cleveland, Cincinnati, Detroit, and Chicago/Milwaukee); Cascadia (Seattle and Portland); northern California (San Francisco Bay Area); and southern California. As these metropolitan areas continue to grow, it is predicted that development patterns will get increasingly dense as older urban cores are revitalized and the suburban land use patterns of the last half of the 20th century become more costly to both inhabit and serve (Barnett, 2007). These assumptions are based upon expected increases in energy costs, an emphasis on transportation and climate change policies that promote more centralized development, and the significant fiscal challenges that local agencies are likely to have in supporting distributed patterns of services. The demographics of these regions are also likely to shift as more affluent younger professionals, aging empty-nesters, and immigrant populations concentrate in the metropolitan cores, a trend that is already advanced in Boston, New York, Chicago, Los Angeles, and San Francisco/Oakland (Nelson and Lang, 2007). In general, our population will be older and more diverse than previous decades, adding to the social vulnerabilities of metropolitan areas. To accommodate the next 100 million people, 70 million housing units will need to be added to the current stock of 125 million; 40 million are likely to be new housing units, while the remaining 30 million are likely to replace damaged or demolished units on existing property (Nelson and Lang, 2007). Also, to accommodate these growing megaeconomies, 100 billion square feet of nonresidential space will likely be added; 30 billion of which is likely to be new
45 square footage, and 70 billion square feet will be rebuilt or replaced (Lang and Nelson, 2007). These statistics were developed before “The Great Recession” slowed housing starts from annual rates of more than 2 million in 2005 and 2006 to 0.5 million in 2009 and 2010, and pushed annual foreÂ� closure rates to more than 3 million (U.S. Census, 2011). The recent recession has also dramatically slowed commercial development and postponed the upgrade of Â�local facilities and infrastructure, much of which was already in sore need of modernization and maintenance before the recent fiscal crisis. To achieve community resilience, now and in the foreseeable future, we must take a more holistic approach to our work as earthquake professionals. With physical resilience as the foundation of our communities’ resilience, we also need to focus on the environmental, economic, social, and institutional resilience of our increasingly dense, complex, and interdependent communities. Also, as past as well as future projections suggest, physical resilience can’t be achieved through expected rates of new construction and redevelopment. It is going to require a whole host of new, innovative engineering solutions, as well as significant and sustained political and professional leadership and will, an array of new financial mechanisms and incentives, and concerted efforts to integrate earthquake resilience into other urban design and social movements. Otherwise, “an earthquake-resilient nation” will remain an idealistic mantra of our profession, and the expected earthquakes of the 21st century will cause unnecessary human, socioeconomic, and physical hardship for the communities they strike. A “Sputnik Moment” in Earthquake Engineering Research In his 2011 State of the Union address, President Obama referred to recent news of technological advances by other nations as this generation of Americans’ “sputnik moment,” and he called for a major national investment “in bioÂ�medical research, information technology, and especially clean energy technology—an investment that will strengthen our security, protect our planet, and create countless new jobs for our people” (White House, 2011). Following the Soviet Union’s launch of the “sputnik” satellite into space in 1957, the United States responded with a major sustained investment in research—most visibly through the establishment of the National Aeronautics and Space Administration (NASA)—and education. The National Defense Education Act of 1958 dramatically increased federal investment in education and made technological innovation and education into national-security issues (Alter, 2011). It is well known that disasters are focusing events for public policy agenda setting, adoption, and change (Â�Birkland, 1997). The September 11, 2001, terrorist attacks put man-made threats at the forefront of disaster policy making, management, and program implementation. September 11 has also been described by some as the major
46 focusing event that significantly expanded the size and scope of the federal government as well as its debt (Stone, 2010). Similarly, Hurricane Katrina has been another focusing event for hazard and disaster management policy and program implementation. To some extent, it has reversed some of the trends started after September 11, but disaster recovery and mitigation have yet to regain their former status as officially preferred disaster policy responses (Waugh, 2006). For earthquake engineering research and seismic policy making, adoption, and change in the United States, the 1971 San Fernando earthquake has been the most recent and Â�salient focusing event. It led to the formation of the Â�California Seismic Safety Commission in 1975 and the passage of the National Earthquake Hazards Reduction Act in 1997 and the formation of NEHRP thereafter (Birkland, 1997). But, was the 1971 earthquake or any other more recent U.S. earthquake a sputnik moment for the United States? The pairing of the 1994 Northridge and 1995 Kobe earthquakes may well have been a sputnik moment for Japan. The tremendous human loss, economic consequences, and, in some Â�cases, surprising causes and levels of damage to structures and infrastructure all contributed to Japan’s major investment in earthquake engineering and disaster management research, education, and policy reform over the past decade. Will we have to wait until a major catastrophic urban earthquake strikes the United States causing unprecedented human and economic losses to have our sputnik moment in earthquake engineering research, practice, and policy reform? If some of the underlying motivations of a sputnik moment stem from shock as well as a sense of being surpassed, then is there any way for our earthquake professional community to better communicate the lessons from Chile versus Haiti and other disasters around the world to compel a more focused policy and investment in earthquake engineering and risk reduction research, education, and action? What can we learn from the biomedical, high-tech, and “green” engineering movements, as examples, which may have recently had, or may currently be a part of sputnik moments, in which policy makers and private investors are motivated to take action in ways that earthquake preparedness has not been able to do with the same growth trajectory and enthusiasm? Ideas for Transformative Earthquake Engineering Research and Solutions The remainder of this paper presents ideas to help foster workshop discussions on transformative earthquake engineering research and on achieving earthquake-resilient communities. It is organized around four dimensions: community resilience, pre-event prediction and planning, design of infrastructure, and post-event response and recovery. Particular emphasis is given to community-level ideas that might utilize the networked facilities and cyberinfrastructure of the George E. Brown, Jr. Network for Earthquake Engineering Simulation (NEES).
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
Community Resilience Drawing upon the research literature of several social science disciplines, Norris et al. (2008) define community resilience as a process linking a network of adaptive capacities in “economic development, social capital, information and communication, and community competence.” To build collective resilience, they recommend that “communities must reduce risk and resource inequities, engage local people in mitigation, create organizational linkages, boost and protect social supports, and plan for not having a plan, which requires flexibility, decision-making skills, and trusted sources of information that function in the face of unknowns” (Norris et al., 2008). To achieve earthquake resilience, we, as earthquake engineering researchers and professionals, need to look beyond earthquakes to other disasters, and even outside of disasters, to understand how our work fits in and how to link our work with other initiatives to build adaptive capacities and incite resiliency-related policy and actions. In 2006, earthquake professionals and public policy advocates joined forces to develop a set of policy recommendations for enhancing the resiliency of existing buildings, new buildings, and lifelines in San Francisco (SPUR, 2009). The San Francisco Planning and Urban Research Association’s (SPUR’s) “Resilient City Initiative” chose to analyze the “expected” earthquake, rather than the “extreme” event, because it is a large event that can reasonably be expected to occur once during the useful life of structures and lifeline systems in the city. It also defined a set of performance goals—as target states of recovery within hours, days, and months following the expected earthquake—in terms of four community clusters: critical response facilities and support systems; emergency housing and support systems; housing and neighborhood infraÂ�structure; and community recovery (SPUR, 2009). Lacking a theoretical model or set of quantifiable measures of community resilience, SPUR relied on expert opinion to set the target states of recovery for San Francisco’s buildings and infrastructure and to assess the current performance status of the cluster elements. For example, SPUR set a target goal to have 95 percent of residences available for “shelter-inplace” by their inhabitants within 24 hours after an expected earthquake; it also estimated that it would take 36 months for the current housing stock to be available for “shelter-in-place” in 95 percent of residences. But, is 95 percent an appropriate target for ensuring an efficient and effective recovery in the city’s housing sector following an expected earthquake? Does San Francisco really need to achieve all the performance targets defined by SPUR to be resilient following an expected earthquake? Which target should be worked on first, second, and so forth? And, given all the competing community needs, when is the most appropriate time to promote an earthquake resiliency policy agenda? There is tremendous need for networked facilities and cyberinfrastructure in support of basic and applied research on community resilience. This includes:
47
APPENDIX B
• Developing an observatory network to measure, monitor, and model the earthquake vulnerability and resilience of communities, including communities’ adaptive capacities across many physical, social, economic, and institutional dimensions. Clearer definitions, metrics, and timescales are needed to establish a baseline of resilience and to measure resilience progress and effectiveness on an urban scale. • Collectively mapping and modeling the individual and organizational motivations to promote earthquake resilience, the feasibility and cost of resilience actions, and the removal of barriers and building of capacities to achieve successful implementation. Community resilience depends in large part on our ability to better link and “sell” physical resilience with environmental, economic, social, and even Â�institutional resilience motivations and causes. • Developing the quantitative models or methods that prioritize and define when public action and subsidy are needed (and how much) to fund seismic rehabilitation of certain building or infrastructure types, groups, or systems that are essential to a community’s resilience capacity versus ones that can be left to market forces, attrition, and private investment to address. • Developing a network of community-based earthquake resiliency pilot projects to apply earthquake engineering research and other knowledge to reduce risk, promote risk awareness, and improve community resilience capacity. Understanding and developing effective, alternative methods and approaches to building local resilience capacity are needed because earthquake-prone communities have varying cultures, knowledge, skills, and institutional capacities. Pre-Event Prediction and Planning To date, much of the pre-event research and practice has focused on estimating the physical damage to individual structures and lifeline systems, creating inventories and scenarios for damage and loss estimation, and preparing government institutions for disaster response. Ideas for “operationalizing” a vision of community-level resiliency include: • Developing more robust earthquake forecasting and scenario tools that address multiple resiliency performance objectives and focus on community-level resilience impacts and outcomes. • Developing more holistic models of individual building risk/resiliency that extend structural simulations and performance testing to integrate information on soil and foundation interaction, non-structural systems, and lifeline systems with the structure and contents information and that model post-disaster building functionality and lifeline dependency and
interdependency and how these affect building functionality, time required to recover various levels of building functionality, and other economic and social resilience factors. • Developing aggregate inventories and models of community or regional risk/resiliency that can be used in mitigation, land use planning, and emergency planning. Local building and planning review processes and emergency management practices need tools to assess the incremental changes in community risk/resiliency over time caused by new construction, redevelopment, and implementation of different mitigation policies and programs. Real estate property valuation and insurance pricing also need better methods to more fully reflect risk and resilience in risk transfer transactions. Within current decision frameworks and practices, redevelopment of a lowdensity, low-rise, but structurally vulnerable neighborhood into a high-density, high-rise development is likely to be viewed as a lowering of earthquake risk and an increase in economic value to the community. But is it really? Tools that more accurately value the aggregation of risk across neighborhoods, incremental changes in community resiliency, effects of aging and densification of the urban environment and accumulation of risk over time, and the dynamics of adaptive capacity of a community post-disaster are needed. • Developing models of the effects of institutional practices and governance on community resilience in terms of the preparedness, recovery, and adaptive capacities. This includes modeling the effects of building and land use planning regulatory regimes, emergency decision-making processes, institutional leadership and improvisational capacities, and postdisaster financing and recovery management policies. Design of Infrastructure Achieving community resilience will require enhanced efforts to upgrade the immense inventory of existing structures and lifelines to be more earthquake-resilient and a broader understanding of resiliency-based performance objectives for building and lifeline design and construction. Ideas include: • Developing enhanced methods for evaluating and retrofitting existing buildings and lifeline systems. These methods need to reliably model the expected responses of existing buildings and lifelines to different levels of ground motions and multiple resiliency performance objectives. Methods need to go beyond estimating the costs to retrofit toward developing more robust models that consider the full range of resiliency benefits and costs of different mitigation
48
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
policy and financing strategies. These alternatives need to think creatively of ways to reuse existing stock, cost-effectively piggy-back on other rehabilitation efforts, and incentivize and ease the burden of retrofitting existing stock. Current and future political challenges are likely to include pressures to preserve historic and cultural integrity, and resistance to invest limited capital resources in seismic rehabilitation projects. Concepts of a federal infrastructure bank might be expanded to include all seismically vulnerable structures and infrastructure, and new public-private financing mechanisms may need to be developed. Mechanisms to effectively communicate the vulnerability of existing structures and lifeline systems to owners, occupants, and policy makers to incite and reward action, such as an earthquake-Â� resilience certification system, also need to be carefully assessed. Sustained efforts to build consensus for standards and actions to evaluate and retrofit existing building and lifeline systems, develop guidelines, and transfer knowledge and technology to building officials, owners, and engineers; utility owners, operators, regulators, and engineers; and other policy and decision makers are also needed. • Advancing performance-based earthquake engineering for buildings, lifelines, geotechnical structures, and hazard defenses. Performance-based engineering needs to take a broader look at the integrated performance of a structure as well as the layers of substructures, lifeline systems, and surrounding community infrastructure that it depends upon. For example, a 50-plus story residential high-rise is, in fact, a neighborhood- or community-vertical, thus making the lifeline conveyance and social resilience of a single structure. Even if the structure is deemed safe following a disaster, lifeline disruptions may impact evacuations and render the structure Â�uninhabitable with sealed windows and lack of elevator service, as examples. • To have resilient communities, we cannot think of building-specific performance only. Communitylevel performance-based engineering models are needed. These may require a systems approach to consider the complex interactions of lifeline systems, critical network vulnerability and dependencies, and dependencies between physical, social, economic, and institutional sectors of a community, and to develop guidelines and “urban-level design standards” for community-level performance. • Making seismic risk reduction and resilience an integral part of many professional efforts to improve the built environment, and building new alliances and coalitions with interest groups working on these goals. This includes the Green Building Council and the Council’s Leadership in Energy and Environ-
mental Design (LEED) program; architects and engineers developing green, adaptive, building “skins,” construction materials, and sensing networks; and urban designers working on sustainable community standards and practices. Current efforts to build new “smart” buildings and cities could potentially benefit from the networked facilities and cyberinfrastructure that the earthquake engineering community has developed in managing and processing the sensing data. In turn, earthquake engineering could potentially assist in helping to develop better and more cost-effective “sensing retrofits” of existing structures and lifeline systems to be “smarter” and to better integrate disaster resilience into the green building and sustainable community standards and practices. In November 2010, the Green Building Council reached a major milestone in its short 10-year life span, having certified more than 1 billion square feet of commercial building space (Koch, 2010). Since it was introduced in 2000, the Council’s LEED program has had more than 36,000 commercial and 38,000 single-family homes participating in the program, of which 7,194 commercial projects and 8,611 homes have been completed and certified as LEED compliant (Koch, 2010). Although the costs for becoming LEED certified may be substantially lower than the costs for enhancing seismic performance, it is not a full explanation for the program’s comparative national and marketing successes. Minimizing damage and reducing the Â�deconstruct/construct cycles of development with higher building performance levels should also be considered as benefits in building valuation. Post-Event Response and Recovery To date, much of the post-event research and practice has focused on estimating the physical damage and economic losses caused by earthquakes and aiding government institutions in disaster response. Ideas for enhancing communitylevel capabilities to rapidly recover from disruption and adapt to changing conditions include: • Creating a more integrated multi-disciplinary network and information management system to capture, distill, integrate, and disseminate information about the geological, structural, institutional, and socioeconomic impacts of earthquakes, as well as post-disaster response and recovery. This includes the creation and maintenance of a repository for postearthquake reconnaissance data. • Developing the next generation of post-disaster damage assessments. Post-disaster safety assessment programs are now well institutionalized in local emergency management and building inspection departments, with legal requirements, procedures,
49
APPENDIX B
and training. The next generation of post-disaster assessments might integrate the sensing networks of “smart” buildings and lifeline systems to make it more quickly possible for emergency responders, safety inspectors, and system operators, as well as residential and commercial building occupants themselves, to understand the post-disaster conditions of buildings or systems and resume occupancy and operation safely or seek alternatives. The next generation of assessments could also take a more holistic view of the disaster impacts and losses, Â�focusing on the economic and social elements as well as the built environment. Just like physical damage assessments, these socioeconomic, or resilience, assessments need to be done quickly after a disaster, and also iteratively so that more-informed, and appropriately timed, program and policy responses can be developed. Such assessments need to look at the disaster-related losses, including the ripple effects (i.e., lost wages, tax revenue, and income); the spectrum of known available resource capital (both public and private wealth and disaster financing resources) for response and recovery; social capital; and the potential unmet needs, funding gaps, and shortfalls, to name a few. • Developing the next-generation emergency response and recovery “dashboard” that uses sensing networks for emergency response and recovery, including impact assessment, resource prioritization and alloÂ� caÂ�tion, and decision making. Research from recent disasters has reported on the use of cell phones, social networking, and Internet activity as a validation of post-disaster human activity. They also caution that sensing networks need to be designed to be passive and part of the act of doing something else, rather than requiring deliberate reporting or post-disaster surveys. They also need to be reasonable and statistically active, culturally appropriate, and conscious of the “digital divide” in different socioeconomic and Â�demographic groups. These systems can also push, and not just pull, information that can be valuable in emergency response management and communication. • Sustained documentation, modeling, and monitoring of emergency response and recovery activities, including the mix of response and recovery activities; multi-organizational and institutional actions, funding, interdependencies, and disconnections that both facilitate and impede recovery; and resiliency outcomes at various levels of community (i.e., household, organizational, neighborhood, and regional Â�levels). This is longitudinal research requiring sustained efforts for 5 to 10 years and possibly even longer, which does not fit well with existing research funding models.
Acknowledgments This paper was developed with the support of the Â� National Research Council and the organizers of Grand Challenges in Earthquake Engineering Research: A Community Workshop. The ideas and opinions presented in this paper are those of the author and also draw upon the work of the National Research Council’s Committee on National Earthquake Â�Resilience—Research, Implementation, and Outreach, of which the author was a member, and its final report (NRC, 2011). In addition, the author acknowledges the work of members of the San Francisco Planning and Urban Research Association’s “Resilient City Initiative.” The author also appreciates the suggestions and detailed review of this paper provided by workshop Co-chair Chris Poland and session moderator Arrietta Chakos; Liesel Ritchie for her sharing of ideas; and Greg Deierlein and Reggie DesRoches (authors of companion papers for the workshop) for their helpful discussion while preparing the paper. Appreciation is also extended to the conference Co-chairs, Greg Fenves and Chris Poland, and the NRC staff for their leadership and efforts in organizing this workshop. References Alter, J. 2011. A script for “Sputnik”: Obama’s State of the Union challenge. Newsweek. January 10. Available at www.newsweek.com/2011/01/05/ obama-sputnik-and-the-state-of-the-union.html. Barnett, J. 2007. Smart growth in a changing world. Planning 73(3):26-29. Birkland, T. 1997. After Disaster: Agenda Setting, Public Policy, and Focusing Events. American Governance and Public Policy. Washington, DC: Georgetown University Press. Fishman, R. 2005. The fifth migration. Journal of the American Planning Association 7(4):357-366. Koch, W. 2010. U.S. Green Building Council certifies 1 billion square feet. USAToday.com. Green House. November 15. Available at www.content. usatoday.com/communities/greenhouse/post/2010/11/green-buildingcouncil-one-billion-square-feet-/1. Lang, R. E., and A. C. Nelson. 2007. The rise of the megapolitans. Â�Planning 73(1):7-12. Nelson, A. C., and R. E. Lang. 2007. The next 100 million. Planning 73(1):4-6. Norris, F. H., S. P. Stevens, B. Pfefferbaum, K. F. Wyche, and R. L. Â�Pfefferbaum. 2008. Community resilience as a metaphor, theory, set of capacities, and strategy for disaster readiness. Dartmouth Medical School, Hanover, New Hampshire. American Journal of Community Psychology 41(1):127-150. NRC (National Research Council). 2011. National Earthquake Â�Resilience— Research, Implementation, and Outreach. Washington, DC: The Â�National Academies Press. Olshansky, R. B., and L. A. Johnson. 2010. Clear as Mud: Planning for the Rebuilding of New Orleans. Washington, DC: American Planning Association. Public Strategies Group. 2011. City of New Orleans: A Transformation Plan for City Government. PSG, March 1. Available at media.nola.com/ politics/other/NOLA%20Transformation%20Plan.pdf. SPUR (San Francisco Planning and Urban Research Association). 2009. The Resilient City: A New Framework for Thinking about Disaster Planning in San Francisco. The Urbanist. Policy Initiative. San Francisco, CA. Available at www.spur.org/policy/the-resilient-city.
50 Stone, D. 2010. Hail to the chiefs: The presidency has grown, and grown and grown, into the most powerful, most impossible job in the world. Newsweek. November 13. U.S. Census Bureau. 2000. Census of Housing. Available at www.census. gov/hhes/www/housing/census/historic/owner.html. U.S. Census Bureau. 2004. Statistical Abstract of the United States. Available at www.census.gov/prod/www/abs/statab2001_2005.html. U.S. Census Bureau. 2011. Census Bureau Home Page. Available at www. census.gov. U.S. Department of Education. 2003. State and County Estimates of Low Literacy. Institute of Education Sciences. Available at nces.ed.gov/naal/ estimates/StateEstimates.aspx.
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH Waugh, W. L. 2006. The political costs of failure in the Katrina and Rita disasters. The Annals of the American Academy of Political and Social Science 604(March):10-25. White House. 2010. National Security Strategy. May 27. Washington, DC. Available at www.whitehouse.gov/sites/default/files/rss_viewer/ national_security_strategy.pdf. White House. 2011. Remarks of President Barack Obama in State of the Union Address—As Prepared for Delivery. January 25. Available at www.whitehouse.gov/the-press-office/2011/01/25/remarks-presidentbarack-obama-state-union-address.
51
APPENDIX B
GRAND CHALLENGES IN LIFELINE EARTHQUAKE ENGINEERING RESEARCH Reginald DesRoches, PhD Professor and Associate Chair, School of Civil & Environmental Engineering Georgia Institute of Technology Summary Lifeline systems (transportation, water, waste disposal, electric power, gas and liquid fuels, and telecommunication) are intricately linked with the economic well-being, security, and social fabric of the communities they serve, and the nation as a whole. The mitigation of earthquake risks for lifeline facilities presents a number of major challenges, primarily because of the vast inventory of facilities, their wide range in scale and spatial distribution, the fact that they are partially or completely buried and are therefore strongly influenced by soil-structure interaction, their increasing interÂ�connectedness, and their aging and deterioration. These challenges will require a new set of research tools and approaches to adequately address them. The increasing access to high-speed computers, inexpensive sensors, new materials, improved remote sensing capabilities, and infrastructure information modeling systems can form the basis for a new paradigm for lifeline earthquake engineering in the areas of pre-event prediction and planning, design of the next-Â�generation lifeline systems, post-event response, and community resilience. Traditional approaches to lifeline earthquake engineering have focused on component-level vulnerability and resilience. However, the next generation of research will also have to consider issues related to the impact of aging and deteriorating infrastructure, Â�sustainability considerations, increasing interdependency, and systemlevel performance. The current generation of the George E. Brown, Jr. Network for Earthquake Engineering Simulation (NEES) was predicated on large-coupled testing equipment and has led to significant progress in our understanding of how lifeline systems perform under earthquake loading. The next generation of NEES can build on this progress by adapting the latest technological advances in other fields, such as wireless sensors, machine vision, remote sensing, and high-performance computing. Introduction: Lifelines—The Backbone of American Competitiveness The United States is served by an increasingly complex array of critical infrastructure systems, sometimes referred to as lifeline systems. For the purposes of the paper, lifeline systems include transportation, water, waste disposal, electric power, gas and liquid fuels, and telecommunication systems. These systems are critical to our economic competitiveness, national security, and overall quality of life.
Water and wastewater systems support population growth, industrial growth, and public health. Power systems provide lighting to homes, schools, and businesses and energize communications. Transportation systems are the backbone of mobility and commerce and connect communities. Telecommunications systems provide connectivity on the local, national, and global scale. Lifeline systems are the basis for producing and delivering goods and services that are key to economic competitiveness, emergency response and recovery, and overall quality of life. Following an earthquake, lifeline systems provide a critical link to communities and individuals, including water for putting out fires, roads for evacuation and repopulation of communities, and connectivity for emergency communications. The resilience of lifeline systems has a direct impact on how quickly a community recovers from a disaster, as well as the resulting direct and indirect losses. Challenges in Lifeline Earthquake Engineering The mitigation of earthquake hazards for lifeline Â� facilities presents a number of major challenges, primarily because of the vast inventory of facilities, their wide range in scale and spatial distribution, the fact that they are partially or completely buried and strongly influenced by interactions with the surrounding soil, and their increasing interÂ�connectedness. Because of their spatial distribution, they often cannot avoid crossing landslide areas, liquefaction zones, or faults (Ha et al., 2010). One of the challenges in the area of lifeline systems, when it comes to testing, modeling, or managing these systems, is the vast range of scales. Testing or modeling of new innovative materials that might go into bridges or pipelines could occur at the nano (10–9 m), micro (10–6 m), or milli (10–3 m) level, while assessment of the transportation network or fuel distribution system occurs at the mega (10 +6 m) scale. Multi-scale models required for lifeline systems involve trade-offs between the detail required for accuracy and the simplification needed for computational efficiency (O’Rourke, 2010). A second challenge related to the assessment of the performance of lifeline systems is that many lifeline systems have a substantial number of pipelines, conduits, and components that are completely below ground (e.g., water supply, gas and liquid fuel, electric power) or partially underground (e.g., bridge or telecommunication tower foundations) and are heavily influenced by soil-structure interaction, surface faulting, and liquefaction. Hence, a distinguishing feature in evaluating the performance of lifelines is establishing a thorough understanding of the complex soil-structure interaction. A third and critical challenge related to lifeline systems is their vast spatial distribution and interdependency between lifeline systems—either by virtue of physical proximity or via operational interaction. Damage to one infrastructure component, such as a water main, can cascade into damage to
52 a surrounding lifeline component, such as electrical or telecommunications cables, because they are often co-located. From an operational perspective, the dependency of lifelines on each other complicates their coupled performance during an earthquake (Duenas-Osorio et al., 2007), as well as their post-event restoration. For example, electrical power networks provide energy for pumping stations and control equipment for transmission and distribution systems for oil and natural gas. A fourth challenge is the aging of lifeline systems. Many lifelines were designed and constructed 50-100 years ago without special attention to earthquake hazards and are deteriorating (ASCE, 2009). Moreover, many lifeline systems have demands placed on them that are much higher than they were originally designed to have. Many lifeline systems are already damaged prior to an earthquake, which increases their vulnerability. Recent Advances in Lifeline Earthquake Engineering The field of lifeline earthquake engineering has experienced significant progress over the past decade. Early studies in lifeline earthquake engineering focused on component behavior and typically used simple system models. They often looked at the effects of earthquakes on the performance of sub-components within each infrastructure system (e.g., columns in a bridge). As more advanced experimental and computational modeling facilities came online via the NEES program, the effects of larger systems (e.g., entire bridge) and coupled soil-structure systems (e.g., pile-supported wharf) were assessed (McCullough et al., 2007; Johnson et al., 2008; Kim et al., 2009). Most recently, advances in modeling and computation have led to the ability to study entire systems (e.g., transportation networks, power networks, etc.), including the local and regional economic impact of earthquake damage to lifeline systems (Kiremidjian et al., 2007; Gedilkli et al., 2008; Padgett et al., 2009; Romero et al., 2010; Shafieezadeh et al., 2011). Transformative Research in Lifeline Earthquake Engineering A new set of research tools is needed to adequately address the critical challenges noted above, namely the vast range in scales, complex mix of soil-structure-equipment systems, interdependencies, and aging and deteriorating lifeline systems. Modeling and managing interdependent systems, such as electric power, water, gas, telecommunications, and transportation systems require testing and simulation capabilities that can accommodate the many geographic and operational interfaces within a network, and among the different networks. The increasing access to high-speed computers, closedform techniques for near-real-time network analysis, inexpensive sensors, new materials, improved remote sensing
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
Â� capabilities, and building or bridge information modeling (BIM or BrIM) systems can form the basis for a new paradigm for lifeline earthquake engineering in the areas of pre-event prediction and planning, design of the nextgeneration lifeline systems, post-event response, and community resilience. Although the current generation of NEES is predicated on large-coupled testing equipment and has led to significant progress in our understanding of how lifeline systems perform under earthquake loading, the next generation of NEES can build on this progress by adapting the latest technological advances in other fields, such as wireless sensors, machine vision, remote sensing, and high-performance computing. In addition, the coupling of seismic risk mitigation with other pressing global needs, such as sustainability, will require a different way of thinking about lifeline earthquake engineering. Sustainability, in this paper, is defined as the ability to meet the needs of current and future generations by being resilient, cost-effective, environmentally viable, and socially equitable. Lifeline systems account for 69 percent of the nation’s total energy use, and more than 50 percent of greenhouse gas emissions are from lifeline systems, so their continued efficient performance is critical for sustainable development (NRC, 2009). Pre-Event Prediction and Planning Because earthquakes are low-probability, high-Â� consequence events, effective planning is critical to making informed decisions given the risk and potential loses. One of the key tools in pre-event planning is the use of seismic risk assessment and loss estimation methodologies, which combines the systematic assessment of regional hazards with infrastructure inventories and vulnerability models through geographic information systems. The performance of lifeline systems is strongly a function of the seismic hazard and the geological conditions on which the lifeline systems are sited. Lifeline systems are strongly affected by the peak ground deformation, which often comes from surface faulting, landslides, and soil liquefaction. Development of approaches to quantitatively predict various ground motion parameters, including peak ground displacement, will be important for understanding the performance of lifeline systems. This quantitative assessment has traditionally been performed using costly and time-consuming approaches that are only typically done on a local scale. The advent of advanced remote sensing products, from air- and spaceborne sensors, now allows for the exploration of land surface parameters (i.e., geology, temperature, moisture content) at different spatial scales, which may lead to new approaches for quantifying soil conditions and properties (Yong et al., 2008). One of the main challenges in regional risk assessment is the lack of reliable and consistent inventory data. Research is needed in finding better ways to acquire data on the vast
APPENDIX B
inventories contained within a lifeline network. Although researchers have effectively deployed remote sensing technologies following natural disasters (such as LiDAR), research is needed in developing ways that these technologies can be effectively used in acquiring inventory data, including physical attributes of different lifeline systems and at different scales. Pre-event planning will require that we learn from past earthquake events. This will require us to vastly improve post-earthquake information acquisition and management. Comprehensive and consistent information on the earthquake hazard, geological conditions and responses, structural damage, and economic and social impacts observed in previous earthquakes are invaluable in planning for future events. This will provide unprecedented information on the performance of individual lifelines but will also provide critical information on the interaction among lifeline systems. A major effort of the future NEES program should be to develop a comprehensive effort among professional, governmental, and academic institutions to systematically collect, share, and archive information on the effects of significant earthquakes, including on the built and natural infrastructures, society, and the economy. The information will need to be stored, presented, and made available in structured electronic data management systems. Moreover, the data management systems should be designed with input from the communities that they are intended to serve. The use of regional seismic risk assessment is key to pre-event planning for various lifeline systems under conditions of uncertainty. For example, detailed information on the performance of the bridges in a transportation network, coupled with traffic flow models, can inform decision �makers on the most critical bridges for retrofit, and which routes would best serve as evacuation routes following an earthquake (Padgett et al., 2010). Significant progress has been made in understanding the seismic performance of lifeline components (e.g., bridges) via component and large-scale testing and analysis; however, much less is known about the operability of these components, and the system as a whole, as a function of various levels of damage. The use of sensors and data management systems would better allow us to develop critical relationships between physical damage, spatio-temporal correlations, and operability. Finally, as infrastructure systems continue to age and deteriorate, it will be necessary to quantify the in situ condition of these systems so that we can properly assess the increased vulnerability under earthquake loading. A dense network of sensors, coupled with advanced prognostic algorithms, will enable the assessment of in situ conditions, which will allow for better predictions of the expected seismic performance (Kim et al., 2006; Lynch and Loh, 2006; Glaser et al., 2007). Performance-Based Design of Lifeline Systems The earthquake performance of a lifeline system is often closely correlated with the performance of a lifeline
53 component (e.g., pipes, bridges, substations). Significant progress has been made in understanding the performance of the lifeline components using the current generation of NEES facilities (Johnson et al., 2008; O’Rouke, 2007; Abdoun et al., 2009; Ivey et al., 2010; Shafieezadeh et al., 2011). However, additional work is needed in designing these systems, considering their role within a larger network, the interdependent nature of lifeline systems, and the trade-offs in terms of cost and safety associated with various design decisions. One critical tool for performing this type of analysis is regional risk assessment programs, such as HAZUS or REDARS. These programs have traditionally been used to assess and quantify risks; however, they can also be the foundation for design of infrastructure systems based on system performance. One key element that goes into these analyses is a fragility or vulnerability curve. Â�Fragility curves are critical not only for comparing the relative vulnerability of different systems, but also for serving as input to cost-benefit studies and life-cycle cost (LCC) analyses. Although cost-benefit analyses are often conducted for scenario events or deterministic analyses, probabilistic costbenefit analyses are more appropriate for evaluation of the anticipated return on investment in a novel high-performance system, by considering the risk associated with damage and cost due to potential seismic damage. Additionally, LCC analyses provide an effective approach for characterizing the lifetime investment in a system. Models often incorporate costs associated with construction, maintenance, upgrade, and at times deconstruction (Frangopol et al., 1997). The LCC models can be enhanced to also include costs associated with lifetime exposure to natural hazards (Chang and Shinozuka, 1996). Such models offer viable approaches for evaluating the relative performance of different structural systems. Given the increased emphasis on sustainability, the next generation of LCC models can also include aspects of environmental impacts (both in terms of materials usage and construction, and deconstruction resulting from earthquake damage) and weigh them against resilience. For example, although greater material investment is often required to make infrastructure systems more resilient, this may make them less sustainable. Conducting this systems-level design will require access to data on both structural parameters (e.g., bridge configuration), environmental, and operational data (such as traffic flows). One research challenge will be how to design our infrastructure systems using an “inverseproblem” paradigm. For example, a goal in design might be to have power and telecommunications restored within four hours of an earthquake event. Using this information as a constraint, the systems (and subsystems) can be designed to achieve these targets. The next generation of BIM or BrIM systems will provide unprecedented information that can be used in the Â�performance-based seismic design community (Â�Holness, 2008). Building information modeling and associated data acquisition sensors (e.g., 3-D scanning and map-
54 ping) and visualization tools (e.g., augmented and virtual reality) cover precise geometry, spatial relationships, light analysis, geographic information, and quantities and properties of building and other infrastructure components. The Â�earthquake-resistant-design community can take advantage of the BrIM or BIM platform to develop and demonstrate design trade-offs that include seismic vulnerability, constructability, costs, schedule, and energy usage. Post-Event Response and Recovery There is an ever-increasing need for quick, yet datadriven response to earthquake disasters. To be most effective, a focused response to an earthquake needs to be initiated and executed within seconds of the event. The challenge for lifeline systems is the need to assess the damage over a wide range of scales. For example, one must rapidly assess the damage to an entire transportation network, to enable rapid determination of emergency routes and to determine where critical resources should be focused. Also, once investigation teams are deployed to the areas of likely damage, advanced tools are necessary to quantify the damage and structural integrity, particularly in cases where the extent of damage is not obvious. Currently, the U.S. Geological Survey (USGS) program ShakeCast (ShapeMap BroadCast) is a fully automated system for delivering specific ShakeMap products to critical Â�users (Wald and Lin, 2007). Its role is primarily in emergency response, loss estimation, and public information. Caltrans uses ShakeCast to set priorities for traffic re-routing, closures, and inspections following a damaging earthquake. The current generation of ShakeCast flags bridges as high, Â�medium-high, medium, or low priority for inspection, primarily based on the expected level of shaking and the system-level performance of bridges. Advancements in bridge modeling and fragility analysis can provide much more informed decision making for emergency response. For example, fragility curves that provide component-level fragility information (e.g., probability of damage to columns, footings, or shear keys) can be much more informative to bridge inspectors and can significantly increase the speed and effectiveness of bridge inspections following an earthquake (Nielson and DesRoches, 2007; Padgett and DesRoches, 2009). Another challenge that needs to be addressed is the assessment of structural integrity, given damage to various components within a lifeline. Current approaches to assessÂ� ing structural integrity are qualitative and often biased by personal experience. Research is needed to find ways to exploit recent advances in sensor technology and/or machine image technology for post-earthquake assessment of structural integrity. Researchers have recently proposed using high-resolution video cameras mounted on first responders’ outfits for determining structural integrity of buildings following a disaster (Zhu and Brilakis, 2010; German et al., 2011). Using the camera, damage inflicted on critical structural
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
elements (in the form of cracks, spalling, bar buckling, etc.) are detected using state-of-the-art recognition techniques. The detected damage is then superÂ�imposed on the detected concrete column element to measure the damage properties (e.g., length, width, and position of crack). This information could be used to query a database of column tests to determine the likely load-carrying capacity of the column and the structural system as a whole. Significant research is needed to better correlate the damage to individual components with the overall structural integrity of the building system. Recent earthquakes in Haiti, Chile, and New Zealand have illustrated the power of remote sensing and have transformed the way that earthquake reconnaissance is performed (Ghosh et al., 2011). Remote sensing was the source of much information in the early days following the earthquake as data providers, such as DigitGlobe and GeoEye, released newly captured imagery with spatial resolutions of up to 50 cm to aid response efforts. Within days of the earthquake, an aerial remote sensing data collection effort was commissioned by the World Bank, in collaboration with ImageCat and the Rochester Institute of Technology. In direct response, the Global Earthquake Observation Catastrophe Assessment Network (GEO-CAN) community was formed to assist in quantifying building damage using the remotely sensed data and to harness online “crowds” of experts, allowing critical damage assessment tasks to be completed rapidly by a distributed network. Such an approach can be adopted for lifeline systems, although challenges would remain for lifelines that are completely or partially buried. Community Resilience The damage to lifeline systems during an earthquake, and the disruption to the vital services that they provide, are critical to the resulting resilience of a community. Resilience refers to the ability of an individual (or community) to respond and recover following a disturbance. It includes those inherent conditions that allow the system to absorb impacts and cope with the event, as well as post-event adaptive processes that facilitate the ability of the system to reorganize, adapt, and learn in the response to the event (Cutter et al., 2008). Researchers have shown that the economic and social impacts from an earthquake are strongly linked to the performance of lifeline systems (Chang et al., 2008). Vulnerability of communities arises from the complex intersection of the built environment, the natural environment, and human systems. Early social science research on community resilience focused on earthquake prediction, forecasting, and warning. This research led to the development of conceptual and empirical models of risk communication and perception, and warning responses. Advances in mapping and modeling the physical vulnerability of a community or region, through GIS and remote sensing technology, significantly improved our understanding of how disasters put communities at risk.
55
APPENDIX B
The recent increase in access to broadband connections, the widespread availability of affordable global positioning systems (i.e., via cell phones), and social media perhaps provide the greatest opportunity to learn from, respond to, and prepare for earthquakes at the community and individual level. These technologies essentially provide the opportunity for individuals to act as sensors (Shade et al., 2010). The potential for tens of thousands to millions of people to monitor and report on the state of damage following an earthquake, as well as the environmental impacts, organization, and human behavior, can help to provide a wealth of information that is useful for better understanding how disasters unfold. More research is needed, however, on the development of tools to collect, synthesize, verify, and redistribute the information in a useful manner. Research Needs for Lifeline Systems The sections of the paper above highlighted the challenges related to the performance of lifeline systems �during earthquakes and opportunities for transformative research on lifeline systems. Below is a summary of the key research needs on lifelines systems in the areas of preearthquake planning, design, post-earthquake response, and community resilience. 1. Site Response Using Remote Sensing Lifeline systems are strongly affected by the peak ground deformation, which often comes from surface faulting, landslides, and soil liquefaction. There is a need to develop approaches to quantitatively predict various ground motion parameters and to reduce the uncertainty associated with these parameters. Moreover, there is a major research need to use advances in remote sensing products, from airand spaceborne sensors, to explore land surface parameters at different spatial scales that can be used to perform quantitative assessment, which has traditionally been performed using costly and time-consuming approaches that are only typically done on a local scale. 2. Inventory Assessment Using Remote Sensing Technologies One of the main barriers to regional risk assessment is the lack of reliable and consistent inventory data. Research is needed to identify better ways to acquire data on the vast inventories contained within lifeline networks. Although researchers have effectively deployed remote sensing technologies following natural disasters (such as LiDAR), research is needed in developing ways that these technologies can be effectively used in acquiring inventory data, including physical attributes of different lifeline systems.
3. Improved Data Management System for Enhancements in Learning from Earthquakes Pre-event planning will require that we learn from past earthquake events. This will require us to vastly improve post-earthquake information acquisition and management. Comprehensive and consistent information on the earthquake hazard, geological conditions and responses, structural damage, and economic and social impacts observed in previous earthquakes is invaluable in planning for future events. This will provide not only unprecedented information on the performance of individual lifelines, but also critical information on the interaction among lifeline systems. A major effort of the future NEES program should be to develop a comprehensive effort among professional, governmental, and academic institutions to systematically collect, share, and archive information on the effects of significant earthquakes, including on the built and natural infrastructures, society, and the economy. The information will need to be stored, presented, and made available in structured electronic data management systems. Moreover, the data management systems should be designed with input from the communities that they are intended to serve. 4. Improved Fragility Relationships for Lifeline Component and Systems Testing of lifeline components has provided important information on fragility relationships. However, more research is needed to develop simulation-based fragility curves, which include component fragility, system fragility, and critical information on functionality, repair time, and repair cost. Such enhanced curves will significantly improve regional seismic risk assessment and form a basis for system-level design of lifelines. 5. Use of Infrastructure Information Modeling Systems for Performance-Based Lifeline Design The next generation of BIM or BrIM systems will provide unprecedented information that can be used in the performance-based seismic design of lifeline systems. The earthquake-resistant-design community can take advantage of the BIM-type platform to perform high-resolution simulation modeling and demonstrate design trade-offs that include seismic vulnerability, constructability, costs, schedule, and energy usage. 6. Pre-Earthquake and Post-Earthquake Condition Assessment Many lifelines were designed and constructed 50-100 years ago, and they are rapidly aging and deteriorating. The recent advances in sensor technology and non-destructive health monitoring methodologies provide significant oppor�
56 tunities for assessing the condition of infrastructure systems in the field. Research is needed in quantifying the deterioration of infrastructure systems and how this deterioration increases the vulnerability of various systems. In addition, there is a need to use sensor technology and/or machine Â�vision technology, coupled with damage detection algorithms and analytical models, to conduct rapid post-earthquake damage assessment and determination of structural integrity. 7. Post-Earthquake Damage Assessment Using Remote Sensing Recent earthquakes in Haiti, Chile, and New Zealand have illustrated the power of remote sensing and have transformed the way that earthquake reconnaissance is performed. Additional research is needed to exploit the use of remote sensing technologies to identify the spatial distribution of damage to lifeline systems and the associated economic and operational impacts. 8. Development of Relationships Between Physical Damage and Operability Although significant progress has been made in understanding the seismic performance of lifeline components via component testing, large-scale testing, and analysis, much less is known about the operability of these components and the system as a whole, as a function of various levels of damage. Research is needed to develop more accurate relationships between component- and system-level damage, and the corresponding functionality, repair cost, and downtime of the lifeline system. The use of sensors and data management systems would better allow us to develop critical relationships between physical damage, spatio-temporal correlations, and operability. 9. Systems-Level Design of Lifeline Systems Significant progress has been made in designing individual lifeline components (e.g., bridges, water mains, substation equipment, etc.); however, additional work is needed in designing these systems, considering their role within a larger network, the interdependent nature of lifeline systems, and the trade-offs in terms of cost and safety associated with various design decisions. One research challenge will be how to design our infrastructure systems using an “inverseproblem” paradigm. For example, a goal in design might be to have power and telecommunications restored within four hours of an earthquake event, considering the relationship between damage and functionality, and the expected interdependency between lifeline systems.
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
10. Improvement in Probabilistic Cost-Benefit Methodologies Research is needed to determine how probabilistic costbenefit analyses can be used to assess anticipated return on investment in new materials, novel high-performance systems, and retrofit. Moreover, research is needed on how lifecycle cost analyses can be used to characterize the lifetime investment in lifeline systems. The models can incorporate costs associated with construction, maintenance, upgrade, and, at times, deconstruction. 11. Sustainability Given the increased emphasis on sustainability, the design of the next generation of lifeline systems must consider both resilience and sustainability. The major research need is to determine how life-cycle cost models can be used to incorporate environmental impacts (both in terms of mateÂ� rials usage and construction, and deconstruction resulting from earthquake damage) and weigh them against resilience. 12. Interdependencies The interdependencies of lifelines complicate their coupled performance during an earthquake. Research is needed to better quantify the correlations between lifeline systems and how to better design the systems when their interactions are considered. 13. Applications of Citizens as Sensors Research is needed on how cell phones and social media can be used to learn from, respond to, and prepare for earthquakes at the community and “individual” level. The potential for tens of thousands to millions of people to monitor and report on the state of damage following an earthquake, as well as the environmental impacts, organization, and Â�human behavior, can help provide a wealth of information that is useful for better understanding how disasters unfold. More research is needed, however, on the development of tools to collect, synthesize, verify, and redistribute the information in a useful and effective manner. Acknowledgments This paper was developed with the support of the National Science Foundation and the National Research Council. The ideas and opinions presented in this paper are those of the author, with considerable input from members of the lifeline earthquake engineering community. The Â�author would like to recognize the following individuals for their valuable comments and suggestions: Tom O’Rourke, M. Saiid Saiidi, Leonardo Duenas-Osorio, Jerry Lynch, Jamie Padgett, Yang Wang, Glenn Rix, Michael Symans, and
APPENDIX B
Mary Lou Zoback. A special thanks to Sharon Wood, Ron Eguchi, and Arrietta Chakos for their detailed review of the manuscript and helpful comments, and Laurie Johnson and Greg Deierlein (who are authoring companion papers) for their helpful discussion while preparing the paper. Finally, I would like to thank the conference Co-chairs, Greg Fenves and Chris Poland, for their leadership in organizing this workshop. References Abdoun, T. H., D. Ha, M. J. O’Rourke, M. D. Symans, T. D. O’Rourke, M. C. Palmer, and H. E. Stewart. 2009. Factors influencing the behavior of buried pipelines subjected to earthquake faulting. Soil Dynamics and Earthquake Engineering 29(3):415-427. ASCE. 2009. Report Card for America’s Infrastructure. American Society of Civil Engineering. Available at www.infrastructurereportcard.org. Chang, S. E., and M. Shinozuka. 1996. Life-cycle cost analysis with natural hazard risk. Journal of Infrastructure Systems 2:118. Cutter, S. L., S. Barnes, M. Berry, C. Burton, E. Evans, E. Tate, and J. Webb. 2008. A place-based model for understanding community resilience to natural disasters. Global Environmental Change 18(4):598-606. Dueñas-Osorio, L., J. I. Craig, and B. J. Goodno. 2007. Seismic response of critical interdependent networks. Earthquake Engineering and Structural Dynamics 36(2):285-306. Frangopol, D. M., K.-Y. Lin, and A. C. Estes. 1997. Life-cycle cost design of deteriorating structures. Journal of Structural Engineering 123:1390-1401. Gedikli, A., M.A. Lav, and A. Yigit. 2008. Seismic Vulnerability of a Natural Gas Pipeline Network. Pipeline Asset Management: Maximizing Performance of Our Pipeline Infrastructure. Proceedings of Pipelines Congress 2008. ASCE Conference Proceedings. DOI: 10.1061/40994(321)77. German, S., I. Brilakis, and R. DesRoches. 2011. Automated detection of exposed reinforcement in post-earthquake safety and structural evaluations. Proceedings of the 2011 ISEC-6 Modern Methods and Advances in Structural Engineering and Construction Conference, June 21-26, 2011, Zurich, Switzerland. Ghosh, S., C. K. Huyck, M. Greene, S. P. Gill, J. Bevington, W. Svekla, R. DesRoches, and R. T. Eguchi. 2011. Crowdsourcing for rapid damage assessment: The Global Earth Observation Catastrophe Assessment Network (GEO-CAN). Earthquake Spectra, In Press, March. Glaser, S. D., H. Li, M. L. Wang, J. Ou, and J. P. Lynch. 2007. Sensor technology innovation for the advancement of structural health monitoring: A strategic program of US-China research for the next decade. Sensor Structures and Systems 3(2):221-244. Ha, D., T. H. Abdoun, M. J. O’Rourke, M. D. Symans, T. D. O’Rourke, M. C. Palmer, and H. E. Stewart. 2010. Earthquake faulting effect on buried pipelines—Case history and centrifuge study. Journal of Earthquake Engineering 14(5):646-669. Holness, G. V. 2008. Building information modeling gaining momentum. ASHRAE Journal (June):28-30. Ivey, L. M., G. J. Rix, S. D. Werner, and A. L. Erera. 2010. A framework for earthquake risk assessment for container ports. Journal of the Transportation Research Board 2166:116-123.
57 Johnson, N., R. T. Ranf, M. S. Saiidi, D. Sanders, and M. Eberhard. 2008. Seismic testing of a two-span reinforced concrete bridge. Journal of Bridge Engineering 13(2):173-182. Kim, J., J. P. Lynch, R. L. Michalowski, R. A. Green, P. G. Mohammed, W. J. Weiss, and A. Bradshaw. 2009. Experimental study on the behavior of segmented buried concrete pipelines subject to ground movements. Â�Proceedings of the International Society for Optical Engineering, Vol. 7294. Kim, J. Y., L. J. Jacobs, J. Qu, and J. W. Littles. 2006. Experimental characterization of fatigue damage in a nickel-base superalloy using nonlinear ultrasonic waves. Journal of the Acoustical Society of America 130(3):1266-1272. Kiremidjian, A., J. Moore, Y. Y. Fan, O. Yazlali, N. Basoz, and M. Williams. 2007. Seismic risk assessment of transportation network systems. Journal of Earthquake Engineering 11(3):371-382. Lynch, J. P., and K. J. Loh. 2006. A summary review of wireless sensors and sensor networks for structural health monitoring. The Shock and Vibration Digest 38(2):91-128. McCullough, N. J., S. E. Dickerson, S. M. Schlechter, and J. C. Boland. 2007. Centrifuge seismic modeling of pile supported wharves. GeoÂ� technical Testing Journal 30(5):349-359. Nielson, B., and R. DesRoches. 2007. Analytical fragility curves for typical highway bridge classes in the central and southeastern United States. Earthquake Spectra 23(3):615-633. NRC (National Research Council). 2009. Sustainable Critical Infrastructure Systems: A Framework for Meeting 21st Century Imperatives. Washington, DC: The National Academies Press. O’Rourke, T. D. 2007. Critical infrastructure, interdependencies, and resilience. National Academy of Engineering Bridge Magazine 37(1). O’Rourke, T. D. 2010. Geohazards and large, geographically distributed systems. Geotechnique 60(7):505-543. Padgett, J. E., and R. DesRoches. 2009. Retrofitted bridge fragility analysis for typical classes of multi-span bridges. Earthquake Spectra 29(1):117-141. Padgett, J. E., K. Dennemann, and J. Ghosh. 2009. Risk-based seismic life-cycle cost-benefit analysis (LCC-B) for bridge retrofit assessment. Structural Safety 32(3):165-173. Padgett, J. E., R. DesRoches, and E. Nilsson. 2010. Regional seismic risk assessment of bridge network in Charleston, South Carolina. Journal of Earthquake Engineering 14(6):918-933. Romero, N., T. D. O’Rourke, L. K. Nozick, and C. R. Davis. 2010. Seismic hazards and water supply performance. Journal of Earthquake Engineering 14(7):1022-1043. Shade, S., G. Lurashci, B. De Longueville, S. Cox, and L. Diaz. 2010. Citizens as Sensors for Crisis Events: Sensor Web Enablement for Volunteered Geographic Information. WebMGS. Shafieezadeh, A., R. DesRoches, G. Rix, and S. Werner. 2011. Seismic performance of pile supported wharf structures considering soil structure interaction in liquefied soils. Earthquake Spectra, In Review. Wald, D., and K.-W. Lin. 2007. USGS ShakeCast. U.S. Geological Survey Fact Sheet 2007-3086. Golden, CO. Yong, A., S. E. Hough, M. J. Abrams, H. M. Cox, C. J. Wills, and G. W. Simila. 2008. Site characterization using integrated imaging analysis methods of satellite data of the Islamabad, Pakistan, region. Bulletin of the Seismological Society of America 98(6):2679-2693. Zhu, Z., and I. Brilakis. 2010. Concrete column recognition in images and videos. Journal of Computing in Civil Engineering 24(6):478-487.
58
EARTHQUAKE ENGINEERING RESEARCH NEEDS IN THE PLANNING, DESIGN, CONSTRUCTION, AND OPERATION OF BUILDINGS Gregory G. Deierlein, PhD, PE John A. Blume Professor of Engineering Stanford University Introduction Past reports on research needs and grand challenges in earthquake engineering have described the importance of building performance to control earthquake losses and lifesafety risks (e.g., EERI, 2003; NRC, 2004). A 2000 HAZUS study estimates a $4.4 billion annualized earthquake loss associated with buildings in the United States (EERI, 2003); the actual losses are likely to be larger with today’s building inventory. Scenario studies of large earthquakes in the United States suggest the losses to be on par or larger than those caused by Hurricane Katrina, with $100 to $200 billion in economic losses and damaged buildings displacing hundreds of thousands of residents and thousands of businesses (Kircher et al., 2006; Jones et al., 2008). Earthquake risks are generally considered to be increasing as the population growth in cities and urban regions is outpacing mitigation measures. In the United States, more than 75 million people live in urban regions with moderate to high seismic hazards (NEHRP, 2008), and this number will continue to climb because of the increasing population and societal pressures toward more dense urban communities. Risks in earthquakethreatened cities are even more pronounced outside of the United States, particularly in developing countries that are experiencing rapid urban growth (Deutsche Bank Research, 2008). It is generally accepted that the most significant earthquake threats affecting buildings in the United States are those associated with (1) casualty risks from collapse in existing buildings that are seismically deficient relative to current building code standards and (2) excessive economic losses, business interruption, and displacement of residents caused by earthquake damage to new and existing buildings. The latter point reflects the fact that current building codes primarily deal with life-safety and do not explicitly address the broader performance factors that can impact communities. With the goal of promoting community resilience to earthquake threats, the recent San Francisco Planning and Urban Research Association (SPUR) study (2009) proposes specific targets for building performance that go beyond basic building code requirements. Specifically, the SPUR study defines five levels of building performance, described in terms of safety and post-earthquake functionality (e.g., safe and operational, safe and useable during repair, etc.). These descriptions of performance highlight important societal needs for maintaining key services (emergency, medical,
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
government, among others) and for sheltering residents in place after a large earthquake. Methods to accurately assess earthquake damage and its impact on continued occupancy and functionality of buildings are essential to implement SPUR’s resilient city vision. Performance-based earthquake engineering provides the means to quantify building performance in terms of (1) collapse and fatality risks, (2) financial losses associated with damage and repairs, and (3) loss of function and recovery time. These performance metrics are intended to inform earthquake risk management and mitigation decisions by building owners, financial/insurance interests, public building code officials, and other stakeholders. The implementation of performance-based engineering requires computational models, criteria, and enabling technologies to (1) simulate building response and performance, (2) design and configure systems with the desired performance, and (3) create building materials, components, and systems that fulfill the design intent. Although there has been significant progress on performance-based methods over the past two decades, continued research is needed to fully achieve and implement the vision of performance-based design for earthquakes. This paper summarizes significant research and development needs for the assessment, design, and creation of earthquake-resilient communities. Although the fundamental concepts of earthquake safety and resiliency are not new, performance-based engineering strategies and methods for addressing the needs are new. Coupled with emerging computational, sensing, and information technologies, the performance-based methods promise to transform the practices of earthquake engineering and risk management. The paper is intentionally focused toward fundamental research and development that fall under the mission of the National Science Foundation (NSF), and it does not attempt to broadly address all of the issues that are important to earthquake risk management. In particular, the paper does not address more applied research and development, technology transfer, and adoption/enforcement of building code standards that fall under the mission of other federal and state agencies. Moreover, the focus in this paper relates specifically to earthquake concerns related to buildings. Broader research needs for earthquake-resilient communities and civil infrastructure are covered in companion workshop papers (DesRoches, 2011; Johnson, 2011); other important scientific research areas, such as geophysics and seismology, are outside the scope of this paper. The paper begins with a summary of research challenges and needs to achieve the vision of resilient communities through performance-based engineering. These needs form the basis for topical research thrust areas that are described in the next section. The paper concludes with brief comments on research facilities and organizations that would be required to conduct the research.
59
APPENDIX B
Research Challenges and Needs for Buildings Research needs and challenges for buildings can be broadly distinguished between those associated with either pre-earthquake planning, design, and construction, or postearthquake response, evaluation, and restoration. Within each category, further distinctions can be made between needs related to (1) new versus existing buildings, (2) individual buildings versus large building inventories, and (3) shortterm (rapid-response) decision making versus longer term planning. Although some research needs are specific to one category, many of them cut across boundaries and pertain to multiple situations. Therefore, in the following discussion, the research challenges and needs are presented in a single group, and comments on the likely applications are discussed under each topic. 1. Simulation of Structural System Response Accurate nonlinear analysis of structural system response, from the initiation of damage to the onset of collapse, is essential to modern performance-based design methods. Analysis of damage initiation is a major factor in assessing post-earthquake functionality and losses associated with repair costs and downtime, and assessment of collapse is a fundamental metric in establishing minimum life-safety requirements for buildings. For example, the FEMA P695 (2009) procedures for evaluating seismic building code design requirements are based on collapse capacities calculated by nonlinear dynamic analysis. Accurate simulation models of damaged structures are likewise important to assess the post-earthquake safety of buildings to aftershocks and to establish requirements for structural repair. In spite of significant advances in nonlinear structural analysis, state-of-the-art methods are still fairly limited in their ability to model nonlinear dynamic response. This is especially true as structures approach collapse, where failure is often triggered by localized concentration of strains, resulting in strength and stiffness degradation that is sensitive to loading history. Examples of such behavior include local buckling and fracture in steel (including both structural steel and steel reinforcement in concrete), shear failures in concrete columns and walls, and connection or splice failures. Current methods to simulate these effects rely heavily on phenomenological models, which rely on empirical calibration and are limited in applicability and/or accuracy by the test specimen sizes, configurations, design parameters, and loading histories considered in the calibration testing. New high-fidelity analyses are desired whose model formulations represent the underlying mechanics and material behavior more directly, such that the models can capture energy dissipation, strength and stiffness degradation, and other effects under any arbitrary loading. The model formulations should incorporate basic material and topology model parameters
and should be validated through large-scale testing of realistic structural components and systems. 2. Comprehensive Assessment of Building Performance Beyond improved structural analysis, research is needed to develop more robust and accurate models to assess complete building performance. Performance-based procedures, such as embodied in ATC 58 (ATC, 2011), provide a fairly comprehensive framework to evaluate building performance. However, current methodologies rely almost entirely on empirical fragility models to assess post-earthquake functionality, damage, and repairs to structural, architectural, electrical, and mechanical building components. Although empirical fragility models offer a practical approach to evaluating component performance, they are inherently constrained by available test data, the types and number of components tested, and the realism of the tests (e.g., representation of boundary conditions, loading, etc.). Moreover, most fragility testing does not provide direct measures of impact assessment (e.g., implications of the damage on other components and systems, restoration and recovery times, etc.). Instead, the impacts are often defined and quantified based on ad hoc judgments and experience. In the near term, component testing and empirical �fragility models will likely continue to play an important role toward implementing performance-based methods. However, looking further into the future, more research should be directed toward developing simulation-based �fragility models, whereby the component behavior is explicitly modeled to calculate component damage and related impacts on functionality, repairs, and recovery. For example, one can envision detailed models of ceiling systems, including lighting, HVAC, and sprinkler piping, where the physical behavior of the overall system, including component damage and functionality, is directly simulated. Such simulations could include direct modeling of damage to the ceiling components, along with cascading damage and restoration, such as simulation of sprinkler piping failures, water damage, and reconstruction operations to repair and restore the facility. Modern building information modeling (BIM) is an important enabling technology to facilitate high-resolution simulation modeling and data management for the wide �variety of structural, architectural, and other building systems and components. 3. Life-Cycle Analysis and Decision Making Life-cycle analysis of economic and other performance metrics is an important tool for performance-based engineering for earthquake risk management and decision making. As with the prior topic (building performance assessment), general frameworks for life-cycle analysis models are fairly well established; however, beyond the basic tools of performance-based assessment, utilization of life-cycle assessment
60 is limited by (1) a lack of accurate models and information on post-earthquake building repair costs and (2) financial and other metrics to quantify the effects of building downtime, displacement of building occupants, and environmental sustainability. Research should focus on filling these needs and to facilitate the integration of life-cycle analysis with earthquake risk mitigation and management. 4. High-Performance Building Systems and Materials Whereas earthquake engineering has traditionally been focused on strengthening and toughening conventional mate� rials and systems to provide minimum life-safety protection, more robust and damage-resistant systems are needed to achieve the vision of earthquake-resilient communities. Seismic base isolation offers perhaps the highest level of seismic protection available today, but it comes with a significant cost premium and other requirements that limit its application. Therefore, research is needed to develop and test other types of high-performance systems, particularly ones that are economically competitive with conventional systems and viable for widespread implementation. Examples of recent developments in damage-resistant seismic systems include (1) self-centering precast concrete and steel framing systems, which employ controlled rocking and elastic posttensioning, (2) moment frame or wall systems that employ high-damping devices, and (3) architectural partitions and other non-structural components that are more resistant to damage from building drifts and accelerations. Given the growth of urban regions and the need for high-density housing and businesses, new earthquake damage-resistant systems for mid- to high-rise residential and office buildings are especially needed. Ideally, these newly developed systems should utilize construction automation, prefabrication, and holistic integration of structural and non-structural components to resist earthquake effects. These needed innovations can �apply to both new buildings and retrofit of existing buildings. 5. Development and Evaluation of Repair Technologies In contrast to the significant research on the design and behavior of new construction and pre-earthquake retro�fit, comparatively less research is available to develop and evaluate common earthquake repair techniques. For example, it is very common to employ epoxy injection to fill earthquake-induced cracks in concrete walls and frames; however, there is comparatively little research to demonstrate the effectiveness of epoxy injection to restore the strength and stiffness of the damaged components. To further support efforts to quickly restore buildings to service, research should focus on (1) technologies to rapidly assess earthquake damage and its effect on building safety to earthquake aftershocks and (2) innovative repair techniques that can be implemented quickly with minimal disruption to building occupants and operations.
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
6. Sensing and Rapid Damage Assessment In further support of the previous need for building repair methods, development of improved sensors and damage assessment methodologies could greatly facilitate post-earthquake response and recovery. Observations from past earthquakes reveal instances where safe buildings are inadvertently closed and taken out of service because of a lack of reliable information on the condition of the structure. Although there is an understandable tendency to err on the conservative side when making decisions about building closure, this conservatism exacerbates problems with displaced residents or loss of important building services. In some instances, overly pessimistic tagging of buildings may even result in buildings being unnecessarily abandoned and demolished. On the other hand, there may also be instances where unsafe buildings remain open and occupied, although these instances are probably less common. In either case, accurate and timely assessment of building conditions would be facilitated through improved sensors and diagnostic tools that can provide immediate feedback as to the building integrity following a significant earthquake. 7. Characterizing Ground Motions and Site/Foundation Response Although free-field ground motions and/or ground motion intensities are commonly used as input for seismic design and analysis, it is generally recognized that the effective input ground motion effects can differ significantly from free-field motions. For example, in stiff short-period buildings, the free-field motions are considered by many engineers to over-estimate the effective input ground motions to the structure. These impressions are supported by the discrepancy between calculated versus observed damage to short-period buildings. Similar trends have been observed in other (longer period) buildings, depending on the building site and foundation conditions. It is hypothesized that the free-field ground motions are reduced by localized deformations in the nearby soil and the soil-foundation interface, but information to confirm this is lacking. Other situations where definition of input ground motions is complicated are in (1) buildings with deep basements, (2) buildings where ground conditions vary considerably across the building site, and (3) dense urban regions where the localized ground motions are influenced by the proximity of closely located buildings, underground structures, and other facilities. Although provisions for soil-structure interaction are available, they are limited in their ability to address situations such as these. Moreover, there is a general lack of well-documented laboratory and/or field data to develop reliable models to characterize soil-foundation-structure interaction and its �effect on input ground motions.
61
APPENDIX B
8. Assessing and Remediating Sites Prone to Ground Deformations Excessive ground deformations are a serious concern for any structure and may result in abandonment of a building site or expensive ground improvement to the site. Moreover, as land becomes scarce in urban regions, there is increasing demand to build structures on marginal sites with soft soils. Over the past 20 years, methods to estimate earthquake ground deformations have improved considerably; however, even the best methods are fairly empirical and have limited resolution to differentiate certain site conditions. As with structural and building systems performance, there is a need to develop simulation-based models that can more accurately calculate expected ground deformations for specific earthquake hazard levels and ground motion. The models should have the capabilities to assess ground deformations at sites with variable soil types (in depth and plan area), and the methods should have sufficient resolution to quantify the effectiveness of ground modification techniques to mitigate the deformations. In addition to developing improved methods to assess ground deformations, research is needed to develop and evaluate new techniques for ground modification, which are more economical and effective than existing methods. 9. Building Benchmarking and Rating With the aim toward improving building codes and promoting effective public policy, research is needed to (1) benchmark the seismic performance of buildings and (2) provide rating methods to make stakeholders more aware of expected building performance and how it can vary between buildings. These benchmarking studies could be done for various purposes, such as to evaluate the performance implied by new buildings that conform to current building codes. Or, studies could be done to benchmark the performance of building types that are predominant in an urban region and to inform policy decisions on seismic safety. To the extent possible, the benchmark studies should be validated (or corroborated) by observed earthquake damage and losses. FEMA P695 (2009) provides a framework for evaluating the collapse safety of buildings, but this procedure relies heavily on judgment to characterize variability in ground motions and model uncertainties; the accuracy of the method is ultimately limited by available nonlinear dynamic analysis models. ATC 58 (ATC, 2011) provides a framework for assessing the more complete response of buildings (casualty risks, direct dollar losses, and facility downtime), but, as noted previously, the performance-assessment techniques and fragility models of ATC 58 are heavily based on empirical evidence and judgment. Therefore, research is needed to develop more robust methods for benchmarking building performance. The impact of benchmarking studies on building codes and policy very much depends on the accuracy (or perceived
accuracy) of the studies. Therefore, it is important that the benchmarking studies be conducted on realistic building inventories, using comprehensive building simulations. Ideally, the benchmark metrics would go beyond collapse safety to incorporate complete building performance metrics (e.g., information on safety and post-earthquake functionality could be interpreted through one of the SPUR building ratings). Modern information technologies would play an important role in storing, accessing, and managing data for the benchmarking studies. Building-specific rating systems to characterize the relative seismic performance between buildings have been proposed as a mechanism that would (1) promote greater awareness of expected earthquake performance of buildings, (2) provide more transparency in seismic design, and (3) encourage more proactive earthquake risk management. The significance of building-specific ratings would be more meaningful if they could be contrasted against comparable ratings for other buildings. Thus, the benchmarking studies of realistic building archetypes would serve an important role in establishing meaningful building ratings. 10. Performance-Based Design and Optimization To date, most of the research on performance-based earthquake engineering has focused on assessing performance, with comparatively less research on how to use more advanced assessment tools to design cost-effective buildings. The implicit presumption is that effective design solutions can be developed by design professionals and then evaluated (checked) for conformance with the desired performance targets. However, given the inherent nonlinearities in the building response and the large number of design �parameters, design optimization for seismic performance can be quite challenging. As performance-assessment technologies are further refined, their practical utilization will require techniques to optimize designs for specific performance targets (e.g., one of the SPUR building resiliency categories) or to minimize life-cycle costs. Therefore, computational design optimization tools are required, whereby a proposed building system can be optimally designed to meet specified performance targets for the lowest life-cycle cost. Design optimization can be applied on a building-specific basis or to archetype building types to develop simplified design provisions for certain classes of structural system or building types. Topical Research Study Areas The research needs outlined above are fairly broad and ambitious, and each will require carefully planned research programs to answer the specific needs. The following is a summary of general research thrust areas that are intended to help describe the necessary experimental, computational, and information technology resources required to respond to the research needs.
62 Probabilistic Framework for Performance-Based Life-Cycle Design and Decision Making Although many of the research needs described in the previous section can be tackled independently, it would be highly desirable to have an overarching performance framework to promote coherency between research projects. This would help ensure, for example, that the computational simulation models, laboratory testing, benchmarking exercises, etc., are coordinated so as to quantify performance criteria in consistent ways that facilitate data and model sharing. To some extent, systematic performance assessment methodologies already exist in the form of HAZUS (Kircher et al., 1997), the PEER methodology (Krawinkler and Miranda, 2004; Moehle and Deierlein, 2004), and ATC 58 (ATC, 2011); however, these frameworks may need to be extended to incorporate simulation-based (in contrast to empirical observation-based) damage, impact, and recovery models in procedures for evaluation and decision making. High-Performance Computational Simulation and Visualization As outlined above, an important innovation in the computational simulation of building performance is an emphasis on fundamental model formulations that better represent the underlying phenomena. This approach is in contrast to more conventional reliance on phenomenological models and simplified fragility models that rely almost exclusively on empirical data and judgment. The proposed simulation models should employ mechanics-based idealizations (finite elements, discrete particle, or other methods) to capture the physical behavior and damage of structural and geotechnical materials and components. Consequences of the damage, associated with repair and recovery operations, could be simulated using construction planning/logistics models. To the extent possible, forward-looking research should apply more fundamental modeling and analysis methods to simulate the performance of non-structural architectural, mechanical, and electrical components and systems. This proposed computational model development is aligned with broader simulation-based engineering and science initiatives at NSF (e.g., Oden et al., 2006; Cummings and Glotzer, 2010; Dyke et al., 2010). As described in these reports, development of fundamental models requires data capture, fusion, and visualization that is similar to the George E. Brown, Jr. Network for Earthquake Engineering Simulation (NEES) vision. The computational demands of the fundamental models will require high-performance petascale (and beyond) computing applications and resources. For some in the earthquake engineering research community, this change in approach will require a culture change. Never� theless, development of fundamental models that require high-performance computing is inevitable and will provide for solutions that are scalable and adaptable to larger and
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
more meaningful problems (e.g., going beyond simulating building components to entire buildings and urban regions). Development and Parameter Calibration of Structural and Geotechnical Models Development of more fundamental (mechanics-based) models for structural and geotechnical materials and components will require high-resolution tests to characterize the underlying mechanics and material behavior that govern behavior. In contrast to traditional experimental research, where large-scale component tests have typically been used to develop “design models” (e.g., nominal strength equations for building codes) and calibrate phenomenological models (e.g., generalized force-deformation models in ASCE 41), the need is for tests that are more directly aimed at development of computational models and calibration of underlying input parameters. Thus, the required testing programs will involve a range of component test specimens to interrogate multi-axial stress and strain states in materials and subÂ� assemblies. These tests will require complementary computational modeling in order to extract appropriate modelÂ�ing parameters. The resulting computational models can then be validated against data from realistic (large-scale) component tests. Consider, for example, reinforced concrete components that experience degradation due to longitudinal reinforcing bar buckling and fracture. These are complex phenomena that require detailed tests to characterize the buckling and fracture behavior more systematically than can be achieved in full-scale column tests. Ultimately, large-scale column tests are needed to validate the models, but large-scale tests alone are not necessarily well suited to model development. Similarly, models of soil-foundation-structure interaction would require detailed soil characterization that builds up to larger system tests. In this regard, the required tests may involve more material and small-scale testing than has been the practice up to now. As noted previously, the resulting analysis models should accurately capture the complete range of behavior from the initiation of damage through to the onset of large inelastic deformations that can trigger collapse. Development and Calibration of Non-Structural Building Component Models Just as with structural and geotechnical components, material and component testing is needed to enable and support the development of simulation-based models for nonstructural building components, such as architectural partitions and finishes, ceiling systems, HVAC, and plumbing and electrical systems and components. For certain components, testing and empirical model development will continue to be the most cost-effective way to develop performance models (e.g., damage and fragility curves). However, where the component behavior is amenable to detailed analysis and/or
63
APPENDIX B
where the distributed systems are too large to test, emphasis should be on testing to characterize features and parameters for computational models. Data Capture and Utilization of Observed Building Performance in Earthquakes Many of the research needs would benefit from more complete data collection to document the performance of real buildings in earthquakes. Because of limitations in resources and training and lack of community leadership, earthquake reconnaissance is often cursory and anecdotal. Although current reconnaissance efforts provide a reasonably good understanding of the “big picture” issues that have arisen in recent earthquakes, the reconnaissance efforts are not generally effective at collecting data and information in ways that support more detailed long-term research. Moreover, sometimes the most important lessons are in buildings that performed well, which are often overlooked during earthquake reconnaissance. The following are suggested as steps toward improved documentation and utilization of data from buildings subjected to earthquakes: (1) protocols should be developed for more consistent procedures for planning and executing post-earthquake reconnaissance; (2) technologies should be developed and made available to facilitate rapid and effective data recording and uploading field observations that include appropriate markers/identifiers for data providence, management, and use; and (3) information technologies should be developed to facilitate storage, management, and use of earthquake observations, including but not limited to (i) photos, videos, and other field observations, (ii) recorded strong motions—both free field and from instrumented structures, (iii) information to classify and model individual buildings and building inventories in earthquake-affected regions, and (iv) information on earthquake losses, impacts on building function and operations, and recovery. Validation Testing of Conventional and Innovative Building Components and Systems Large-scale testing will continue to be important for validating simulation models of both conventional and new building components and systems. However, in contrast to past practice where tests are often run before detailed models have been developed, greater emphasis should be placed on model validation, where detailed analysis models are developed and scrutinized prior to the large-scale tests. Predictive analysis results should reflect the uncertainties in nonlinear behavior and analysis, where the uncertainties are built into the analysis ahead of the test, rather than being rationalized after the test to explain discrepancies between the calculated and measured data. Large-scale validation tests should be planned and developed with significant input and involvement of the research community so as to (1) make sure that
the tests capture the most relevant and important behavioral effects and (2) engage the broader community in making effective use of data from expensive large-scale tests. Research funding models may need to be revised to support this sort of involvement. Development and Validation of Sensors and Damage Models The rapid pace of technological advancements in sensors, wireless communication, and digital information technologies over the past 20 years offer unprecedented opportunities for collecting extensive data from laboratory tests, real buildings, and distributed inventories of buildings. As with many new technologies, much of the previous research on sensing and health monitoring has been on development of sensors, signal processing, and stand-alone damage detection algorithms. Looking ahead, research programs should focus on ways to integrate sensing technologies with (1) computational model updating and validation, (2) interpretation of observed building performance in earthquakes, and (3) rapid post-earthquake assessment of buildings. Given the rapid proliferation of high-resolution imaging (still images and video), there are important research opportunities to investigate ways of automating (or semi-automating) the interpretation and use of image data. Fusion of Inventory Data with High-Fidelity Simulations and Visualization for Benchmarking, Building Rating, and CostBenefit Studies The needs for building benchmarking, seismic rating systems and cost-benefit studies offer ideal testbeds to apply high-performance computing to practical research needs. Such studies would promote close collaboration between researchers in high-performance computing, high-fidelity computational modeling, and earthquake risk assessment, management, and decision making. The studies would help promote realism into the research, which would ultimately lead to greater impacts on building design, adoption of new technologies, and development of policies for seismic risk mitigation. Implications on Research Facilities The research needs and thrust areas described above will require unprecedented coordination and data fusion between high-fidelity computational simulations, building inventory descriptions and information models, laboratory tests, and observations/measurements of building performance during earthquakes. Although large-scale laboratory testing will continue to be a critically important component of earthquake engineering research, increased emphasis should be placed on computational model development and physical testing that is in direct support of its development. As noted previously, the needs for physical testing are not all large
64 scale. In fact, in some cases, the most critical needs are for testing facilities and instruments to characterize material behavior and mechanics at small (micro) scales, e.g., characterizing detailed nonlinear behavior, damage, and failure of structural materials including concrete, steel, soils, wood, composite fibers, etc. At the other extreme, greater emphasis should be placed on methods to measure and collect data from field conditions, including both planned field experiments (including building demolition) and unplanned events (earthquakes and other extreme loadings). Given the breadth and depth of the research needs, many of the needs may best be addressed through research center (or center-like) programs. Many significant facilities are �already in place for large-scale testing, and these will need to be maintained. Although the United States has some remarkable high-performance computing facilities available, these tend to be underutilized for earthquake engineering research. Therefore, there is a critical need for a more concerted effort on simulation-based high-performance computing research, which involves development and validation of (1) models, (2) improved computational algorithms for large models, and (3) tools to facilitate management and integration of massive test and analysis databases. Otherwise, given the current state of earthquake engineering research, it would be difficult for any individual researcher (or small group of researchers) to assemble the critical mass of computational simulation expertise and resources to make the types of transformative changes that are necessary to address the research needs. Acknowledgments This paper was developed with the support and invitation of the organizers of Grand Challenges in Earthquake Engineering Research: A Community Workshop. The ideas and suggestions presented in this paper were developed by the author with input from many members of the earthquake engineering community. Although it is impossible to acknowledge all of the individuals who contributed to the ideas in the paper, the author would like to recognize the following individuals as making specific suggestions that are reflected in the paper: R. Boulanger, G. Brandow, M. Comerio, C. Comartin, G. Fenves, A, Filiatrault, J. Hajjar, R. Hamburger, J. Heintz, W. Holmes, C. Kircher, H. Krawinkler, E. Miranda, J. Moehle, J. Osteraas, C. Poland, K. Porter, J. Ramirez, R. Sabelli, S. Wood. The author would also like to acknowledge helpful discussions with L. Johnson and R. DesRoches, who are authoring companion papers for the workshop. References ATC (Applied Technology Council). 2011. Guidelines for Seismic Performance Assessment of Buildings. ATC Project 58, 75% Draft Guidelines. Redwood City, CA.
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH Cummings, P. T., and S. C. Glotzer. 2010. Inventing a New America through Discovery and Innovation in Science, Engineering and Medicine: A Â�Vision for Research and Development in Simulation-Based Engineering and Science in the Next Decade. Baltimore, MD: World Technology Evaluation Center (WTEC). Available at www.wtec.org/sbes-vision/ RDW-color-FINAL-04.22.10.pdf. DesRoches, R. 2011. Grand challenges in lifeline earthquake engineering research. In Grand Challenges in Earthquake Engineering Research: A Community Workshop. Washington, DC: The National Academies Press. Deutsche Bank Research. 2008. Megacities: Boundless Growth, 18 pp. Available at www.dbresearch.com/PROD/DBR_INTERNET_EN-PROD/ PROD0000000000222116.pdf. Dyke, S. J., B. Stojadinovic, P. Arduino, M. Garlock, N. Luco, J. Rairez, and S. Yim. 2010. Vision 2020: An Open Space Technology Workshop on the Future of Earthquake Engineering. A Report on the NSF-Funded Workshop, January 25-26, 2010, St. Louis, MO. Available at nees.org/ resources/1637/download/Vision_2020__Final_Report.pdf. EERI (Earthquake Engineering Research Institute). 2003. Securing Society Against Catastrophic Earthquake Losses: A Research and Outreach Plan in Earthquake Engineering. Oakland, CA. Available at www.eeri.org/ cds_publications/securing_society.pdf. FEMA (Federal Emergency Management Agency). 2009. Quantification of Building System Performance Factors. FEMA P-695. Prepared by the Applied Technology Council, Redwood City, CA. Johnson, L. A. 2011. Transformative earthquake engineering research and solutions for achieving earthquake resilient communities. In Grand Challenges in Earthquake Engineering Research: A Community Workshop. Washington, DC: The National Academies Press. Jones, L. M., R. Bernknopf, D. Cox, J. Goltz, K. Hudnut, D. Mileti, S. Perry, D. Ponti, K. Porter, M. Reichle, H. Seligson, K. Shoaf, J. Treiman, and A. Wein. 2008. The ShakeOut Scenario. USGS Open File Report 20081150. Reston, VA. Kircher, C. A., R. K. Reitherman, R. V. Whitman, and C. Arnold. 1997. Estimation of earthquake losses to buildings. Earthquake Spectra 13(4):703-720. Kircher, C. A., H. A. Seligson, J. Bouabid, and G. C. Morrow. 2006. When the big one strikes again—Estimated losses due to a repeat of the 1906 San Francisco earthquake. Earthquake Spectra 22(S2):S297-S339. Krawinkler, H., and E. Miranda. 2004. “Performance-Based Earthquake Engineering,” Chapter 9 in Earthquake Engineering from Engineering Seismology to Performance-Based Engineering, Y. Bozorgnia and V. V. Bertero, eds. Boca Raton, FL: CRC Press. Moehle, J., and G.G. Deierlein. 2004. A Framework Methodology for Performance-Based Earthquake Engineering. Proceedings of the 13th World Conference on Earthquake Engineering, Vancouver, B.C., Canada, on CD-ROM. NEHRP (National Earthquake Hazards Reduction Program). 2008. Strategic Plan for the National Earthquake Hazards Reduction Program: Fiscal Years 2009-2013. October. Available at nehrp.gov/pdf/Â�strategic_ plan_2008.pdf. NRC (National Research Council). 2004. Preventing Earthquake Disasters: The Grand Challenge in Earthquake Engineering. Washington, DC: The National Academies Press. Oden, J. T., T. Belytschko, J. Fish, T. J. R. Hughes, C. Johnson, D. Keyes, A. Laub, L. Petzold, D. Srolovitz, and A. Yip. 2006. Simulation-Based Engineering Science: Revolutionizing Engineering Science through Simulation. Report of the National Science Foundation Blue Ribbon Panel on Simulation-Based Engineering Science, 88 pp. Available at www.nsf.gov/pubs/reports/sbes_final_report.pdf. SPUR (San Francisco Planning and Urban Research Association). 2009. The Resilient City: A New Framework for Thinking about Disaster Planning in San Francisco. The Urbanist Policy Initiative. San Francisco, CA. Available at www.spur.org/policy/the-resilient-city.
65
APPENDIX B
CYBERINFRASTRUCTURE-INTENSIVE APPROACHES TO GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH James D. Myers Director, Computational Center for Nanotechnology Innovations Rensselaer Polytechnic Institute Introduction In just the past decade, the power of the largest supercomputers has increased by more than a factor of a thousand. The cost per bit of data storage has decreased by a similar factor. We can buy personal cell phones today that feature Internet connections rivaling those of entire universities of the year 2000. Further, it is clear that such trends in increasing performance and decreasing price for performance ratios will continue in the coming decade. Such dramatic, ongoing change brings both opportunities and challenges to the earthquake engineering community. Research can be done faster, more cost-effectively, on larger systems, and with higher fidelity “simply” by applying the latest technology. However, technological advances bring “nonlinear” opportunities— when, for example, automated sensing becomes cheaper than manual measurement, when simulation becomes more accurate than experiments, when discovering and using existing data becomes simpler than generating new data. Realizing this type of opportunity involves significant changes to “business as usual” along with technology adoption—from the development of new research techniques to additional coordination required for data integration to support for new collaborations and career paths. The earthquake engineering community has made significant scientific and engineering advances through the use of cyberinfrastructure over the past decade: large networks of seismic stations generate ground motion data that are integrated and stored as a long-term community resource; strong wall and shake table experiments record thousands of data channels and video to provide very high resolution information about structural performance; simulation of structural behavior has become a tool for interpreting experiments and extrapolating from them even to the point of becoming part of hybrid experiments; geospatial modeling tools have grown to encompass socioeconomic effects, inter-network interactions, and decision support capabilities; and the communityscale sharing of data, tools, and equipment has accelerated community development and adoption of new techniques. To continue such advancements in the next decade, and to capitalize on emerging cyberinfrastructure capabilities, in pursuit of earthquake engineering grand challenges, it will be important to understand the current state and promising research directions in cyberinfrastructure. Although there are directions in which exponential increases in computing capabilities will continue, there are others, such as the raw clock
speed of individual CPUs, where progress has effectively stopped. There are also areas where synergies from progress on multiple fronts are enabling dramatic new possibilities. Indeed, as this paper is being written, a computer challenged human supremacy at the game of Jeopardy! and won, opening an era in which computers will actively aid researchers in applying reference information to their work. Cyberinfrastructure: A Working Definition There are a number of connected challenges in providing a succinct overview of cyberinfrastructure and its potential. Cyberinfrastructure is broad, both in terms of the underlying suite of technologies and in the potential areas of application. It includes the hardware for sensing, networking, computation, data management, and visualization, as well as the software required to turn these technologies into capabilities—e.g., for automated data collection, modeling, data analysis, and group and community-scale collaboration. Most definitions of cyberinfrastructure, including those from the National Science Foundation (Atkins et al., 2003; NSF, 2007), also recognize its “socio-technical” nature—using it effectively requires changes in practices and culture, shifts in responsibilities between organizations, and even the development of new career paths. In this sense, the definition quickly becomes “everything remotely related to information technology and its use.” However, the definition can be constrained somewhat by restricting it to true infrastructure (Edwards et al., 2007)— areas where economies of scale, coordination challenges, and/or transformative potential argue for shared provisioning, support for cross-disciplinary collaborations, and the creation of common middleware and standards that simplify and guide further development. These considerations naturally refocus the definition in areas of rapid change and where there is significant expected value and broad demand. Cyberinfrastructure development and deployment can thus be distinguished from curiosity-driven research through a requirement for a direct connection to domain (i.e., earthquake engineering) problems, a clear argument for its role in solving them, and defined metrics for its success (Berman et al., 2006). Although this definition is still not fully prescriptive, it does provide a useful framework for the following discussion. The Near Future of Cyberinfrastructure Although predicting the exact future of cyberinfrastructure is difficult, getting a sense of the future landscape is less so. The roadmaps of manufacturers indicate that trends such as the increasing computational capacity of supercomputers will continue for much of the decade and deliver factors of at least 100 to 1,000 times more performance than today. Power efficiency and density will also increase, meaning that desktop and rack-scale systems will also see dramatic
66 performance improvements. However, much of that power will come from increases in parallelism, either from an increase in the number of CPUs integrated together or via increased parallelism within chips, as is already occurring through the use of general purpose graphics processing units—video cards—and their hundreds of parallel pipelines. The corollary here is that software that is not engineered or reengineered to leverage that parallelism will see minimal gains—a dramatic difference from the past decade. Data storage is also increasing rapidly in capacity and decreasing in price. However, an often overlooked fact is that single disk read/write speeds have not kept pace—the average time to retrieve a randomly located byte on a disk has decreased by less than a factor of 10 since the 1970s. This has opened a large—many orders of magnitude—gap between raw compute performance and the ability to compute effectively on large datasets, which is now receiving significant attention under the banner of “data-intensive” computing. Hardware vendors are devoting significant effort to adding fast data caches (multiple levels) to chips, creating solid state (flash memory) disks and caches, and distributing data across many disks for faster access. Innovation is also occurring in file systems and database software, as well as processing software, which is enabling massively parallel processing of data and statistical analysis across large datasets. Early adopters of data-intensive techniques include Â�Internet search engines and processors of text in general (e.g., books as well as web pages), in part because of the massive amount of text available. However, the use of distributed sensors, images and video feeds, and high-throughput analysis techniques are rapidly increasing the need for similar capabilities in scientific research. The size and cost of sensors, the bandwidth available to wireless and mobile sensor platforms, the resolution of cameras, and the size and resolution of displays are all benefitting from the advances in computer chip manufacturing and commodity/consumer applications, resulting again in orders of magnitude of change. In synergy with hardware advances, software advances are occurring, and the combination can be stunning. It has become possible to automatically integrate photos (e.g., tourist photos from Flickr) to reconstruct an underlying 3-D model of the scene (e.g., of the Statue of Liberty). Cars, planes, quad-copters, and submersibles can pilot themselves. Grids can provide scheduled access to tens of thousands of interconnected processors, while clouds can provide truly on-demand access to commodity clusters or to applications and services directly. Computers can automatically monitor thousands of video streams and identify features and events, and they can read the literature and databases and answer questions in natural language. Empowering Individuals to Address Grand Challenges Although it would be entertaining to continue to review the existing and emerging wonders of computing, the goal
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
of planning for cyberinfrastructure-intensive research in earthquake engineering is better served by returning to the definition of cyberinfrastructure and discussing the potential for these technology advances to further progress on grand challenges. The real questions are: What can several orders of magnitude of increases in compute power, data sizes, and sensor density, combined with a rapidly increasing capability to focus those resources on demand and to automate larger and more complex tasks, enable, and what is required beyond the raw technologies to realize those new opportunities? To begin, consider first the ways that increased compute power can be used. In modeling physical processes, 3 to 4 orders of magnitude can provide (only!) a factor of 10 in spatial resolution in three dimensions and time, which would be useful if, and only if, such increases improve fidelity. Or one could model structures 10 times larger at the same resolution. In domains where stochastic and/or chaotic processes occur, the best use of additional power can be to run ensembles of simulations to map the uncertainties in initial measurements to a range of outcomes. Combined with advances in adaptive meshing and multi-scale/multi-domain modeling, processing power can also be expended to do very fine-grained modeling or to use more advanced models in small areas, e.g., where bending is occurring or cracks are developing. The growing capabilities for data generation, storage, analysis, and visualization can be analyzed in a similar way. Increased resolution and/or experiments on larger or more complex samples become practical. Statistical properties can also be computed (assuming sufficient samples and/ or non-destructive testing) and compared with statistical simulations or with reconnaissance data. Even more interestingly, massive amounts of cheap data can substitute for more costly measurements. Aerial photography is already used to infer inventory information and reduce the need for people to physically categorize buildings. One could also imagine dense arrays of randomly located sensors replacing handplaced ones, with positions calibrated through use of multiple image/video feeds. Advances in sensors themselves—to auto-calibrate, to dynamically adjust measurement rate—and for mobile sensor platforms to enable dynamic in-depth observations when and where needed will enable “stretching” within the data budget analogously to the way adaptive and multi-scale modeling can stretch the computational budget to enable more valuable results at lower costs. Implicit in the use of such high-throughput techniques is the challenge of managing more measurement streams and more datasets. Folder/file hierarchical names become increasingly cumbersome as the number of files increases. Further, any manual actions required to find, select, or convert files, or to extract specific data from files, or to transfer files between machines, become bottlenecks. Fortunately, tools for automatically capturing descriptive information about data as well as its provenance (data history) and to allow subsequent human annotation are emerging. Such an infrastructure then opens a range of opportunities—use of
APPENDIX B
provenance for debugging and automating subsequent analyses, automation of reporting and data publication activities, etc. Given the management overheads in running sensor networks and equipment sites at the current scale, the potential to actually reduce those overheads rather than having them scale with, for example, the number of sensors could become critical in maintaining/increasing productivity. Increasing amounts of data, whether from experiments or simulations, raise new challenges beyond those of acquisition and storage. Understanding large amounts of data, and/ or data representing a large number of dimensions, cannot be done by manual inspection and manipulation of the data. Automated detection of features and of events of interest and automated analysis of their correlations and evolution becomes a necessity. (Minimally, feature and event detection are required to provide navigational guides for manual exploration of the data.) As data volumes grow, algorithms to detect known features (e.g., cracks) will be supplemented by “machine-learning” algorithms that detect and report correlations that may or may not have been expected (e.g., if two sets of sensors suddenly have less correlation in their measurements, then it might indicate a broken joint or loose mounting that may or may not have been anticipated by the researcher). The potential to automate the detection of unanticipated features is at the heart of arguments for a “Fourth Paradigm” of research complementing experiment, theory, and simulation. As discussed in previous sections, the raw capabilities to acquire, process, and store increased data volumes have advanced tremendously and will continue to do so. There have also been significant technical advances and cost reductions in the areas of large displays, stereoscopic 3-D displays, and remote visualization over the past few years, but significant challenges remain in effectively using such hardware and, for example, creating capabilities for feature-based navigation, analysis, and display on such devices. For example, tiled displays now exist that provide more than 300 million pixels. Smaller systems, ones that still match the resolution of the human eye (~15 million pixels), can now run un-modified applications (web browsers, Excel, Matlab) using only a single computer (versus the visualization cluster and specialized software required to run larger displays). However, using either class of displays for something more than showing a single large image or massive arrays of sensor traces quickly becomes a custom development effort. Over the next few years though, it should become possible to trigger feature detection algorithms to run automatically as data are stored and for feature-based navigation tools to not only show synchronized display of multiple inputs but also to select which inputs of the hundreds or thousands available are the best to look at to understand a given feature. Direct visualization of the distribution of values across an ensemble of experiments, and/or visualization of the deviations between experimental and simulated results, are also likely to be capabilities that can be quickly configured and automated.
67 Empowering the Community to Address Grand Challenges As significant as the direct benefits of cyberinfrastructure are on an individual’s ability to increase the scope, scale, and speed of his or her research, the community-scale impacts, in terms of increased capacity to address grand challenges, are likely to be greater. For grand challenges, reducing the time-to-solution requires reducing the time required for problems to be understood, for new ideas to be implemented, for successful techniques to be reported and widely disseminated, and for new experimental results to become reference data and influence theories, models, and practice—and vice versa for information to flow in other directions. Cyberinfrastructure, with orders-of-magnitude increases in raw performance, can have dramatic impacts throughout this web of processes. The earthquake engineering community is well versed in the core technologies that enable community Â�coordination— distributed meeting software (e.g., videoÂ�conferencing, shared computer displays), asynchronous collaboration and networking tools (e.g., email, wikis, blogs, Facebook, Twitter), equipment sharing, community databases, remote modeling and simulation services, search engines, and digital libraries, etc. Performance and the price/performance ratio will continue to improve in these areas, driven by the computing trends previously discussed and similar rapid increases in available network bandwidth. However, the more important trend will be automation of the information exchange across the community. Imagine: • sending a model to observe an experiment rather than observing it yourself and receiving alerts when the experiment and model results diverge; • shifting from manually searching the Internet and databases for relevant results to receiving automated notification as new results appear; • shifting from notices to automatic incorporation of such results into the calibration and validation suites of computational models. (Indeed, services that can automatically assess whether new results are consistent or inconsistent with existing reference information, as well as services to dynamically update reference information based on new results, have already been developed in fields such as chemistry [thermodynamics]); • combining such active reference services with domain computational models to enable goal-driven research, e.g., being able to plug individual models and datasets into an open framework to immediately assess their import in achieving regional resilience. In chemistry, such combined capabilities are expected to, for example, help identify which thermochemistry experiments would have the most impact in reducing the uncertainties and errors in modeling engine
68
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
performance. In earthquake engineering, a similar approach might potentially be used to identify which experiments and what additional observational data would most directly address the largest uncertainties in the design of new structures and the estimation of earthquake impacts; or, • simply asking for the evidence in the literature for and against a given theory rather than reading all the papers with relevant keywords. Although such speculation is perhaps utopian, it is hard to argue away the significance of the incremental progress that will be made. The deliberations of expert groups will be informed by automated test of data quality and mathematical “meta”-analysis of the entire corpus of experiments. Data and models will be available as services (“in the cloud”), and sensitivity analyses will inform experiment selection and design. It will become possible to identify specific claims being made in the literature and to retrieve the data and analysis details required to assess them. Further, as data, models, and metadata become more accessible, and as the computational resources available to individuals grows, small groups will tackle and succeed at tasks that would require unsustainable levels of community coordination today. Challenges in Realizing the Potential of Cyberinfrastructure Adoption of information systems and cyberinfrastructure in academia and business has often had what can tactfully be called mixed success. Looked at over the long term, progress is unmistakable. Over shorter timescales and at the scale of individual projects, it is much harder to discern what has and hasn’t worked and to assess how well money and time are being spent. There is growing understanding, however, some from the experiences of the earthquake engineering community, of the nature of the challenges involved and in best practices. At their core, the major challenges are all related to the unprecedented and almost incomprehensible rate of technological advancement. By the time a non-IT specialist can acquire skills in a given technology, it becomes obsolete. Work practices, based on implicit assumptions about the relative difficulties of different tasks, become inefficient or ineffective as technological change invalidates those assumptions. Central coordination mechanisms that can be highly effective in disseminating best practices fail as small groups, empowered by technology, create disruptive innovations. Thus, in addition to the direct challenges of creating and deploying useful technologies, the pace of change creates “metachallenges” to our traditional practices of development and deployment. Fortunately, across the efforts developing and deploying cyberinfrastructure, there has been a concerted effort to address these meta-challenges and to identify social, managerial, and technical practices that are robust in the
face of rapid change (NEES, 2007; Faniel, 2009; Ribes and Finholt, 2009; Sustain, 2009). Placing such general best practices in the context of the proceeding sections can help make them concrete. For example, the shift from faster clock speeds to massive parallelism as the mechanism for increasing performance during the next decade means that graduate students without parallel programming skills will not be able to realize the value of the emerging technology. Explicit support for cross-training, collaborative projects pairing computational scientists and domain researchers, and, most scalably, training in the use of programming frameworks and libraries that embed parallel expertise in their design will all be critical in capitalizing on the raw performance increases. At the project level, there is growing recognition that user-centered, experiment-driven, iterative development practices are most effective in supporting co-evolution of technology and work practices. Central to such methodologies are partnerships in which the end goal of enabling new research is explicit and the design space of solutions includes both technologies and new practices. Such a process targeting the need to compare experimental and simulated results might first address the data conversions necessary to perform experiments and simulations on the same system followed by enhancements that would allow simulations to be run in parallel with experiments and that would enable direct visualiÂ�zation of the differences between them. Such technical changes would be paralleled by changes in practice where simulations might be run before experiments to improve their design, or where adaptive sensing or steering of the experiment might provide additional insight when differences are seen. Providing incremental capabilities that deliver value and thereby affect practice improve efficiency compared to monolithic, pre-planned approaches. At the scale of communities, emerging best practice recognizes that true infrastructure is not a system, but a system of systems, which has consequences for design and management. Designs appropriate for infrastructure support “innovation at the edges” and focus on standardizing common functionality. Well-known examples include Internet routing, which does not constrain what is sent across the Internet, and HTTP/HTML, which defined formatting and linking rules but did not constrain the content of web pages. Cyberinfrastructure designs such as web services, workflow, content management, global identifiers and vocabularies (the basis of the semantic web), and separable authentication mechanisms (enabling single-sign-on) work analogously, providing simple, best practice means for addressing common problems without constraining future innovation. At the level of grand challenges, where many independent organizations must coordinate as peers, such designs are critical (ALA, 2008). Management structures that encourage such designs—independent processes for defining interfaces and for implementation within the defined framework, early definition of interfaces and competitive funding of functionality
69
APPENDIX B
written to those interfaces, and inter-project and interdisciplinary communication to identify functionality ripe for standardization, for example—are a critical complement. Such non-technical concerns extend to the implementation of mechanisms to recognize inter-disciplinary work and to support career paths for those who cross such boundaries.
lenges in Engineering Research: A Community Workshop. The author wishes to acknowledge contributions by many members of the Network for Earthquake Engineering Simulation, the Mid-America Earthquake Center, and the large earthquake engineering community to the concepts presented here through many informal discussions and formal planning exercises over the past decade.
Conclusions The technical advances that have made cyberinfrastructure-centric approaches to scientific and engineering endeavors a major driver of overall scientific progress will continue and broaden in the next decade. Progress will bring more computational and data management capabilities to individual researchers, while the decreasing cost of those resources will allow their use more broadly to automate manual processes throughout the scientific lifecycle. Application of cyberinfrastructure to coordinate and accelerate communityscale processes will further increase the earthquake engineering community’s collective ability to tackle grand challenge issues. Although this paper has touched on many aspects of cyberinfrastructure and highlighted a number of potential uses, there are both technologies and application areas (e.g., most glaringly, education and outreach) that have not been discussed because of space limitations. However, the core conclusion from this paper is not about any specific application of cyberinfrastructure. Rather, the conclusions are that the underlying trends, and even relatively straightforward analysis of the potential applications, make it clear that further investment in cyberinfrastructure should be very profitable in terms of impact on grand challenge agendas and that experience over the previous decade provides clear guidance concerning what will be required to realize that value. Acknowledgments This paper was developed with the support of the Â�National Research Council and the organizers of Grand Chal-
References ALA (American Lifelines Alliance). 2008. Post-Earthquake Information Systems (PIMS) Scoping Study. Washington, DC: National Institute of Building Sciences. Atkins, D. E., K. K. Droegemeier, S. I. Feldman, H. Garcia-Molina, M. L. Klein, D. G. Messerschmitt, P. Messina, J. P. Ostriker, and M. H. Wright. 2003. Revolutionizing science and engineering through cyberinfrastructure: Report of the National Science Foundation Blue-Ribbon Advisory Panel on Cyberinfrastructure. Berman, F., J. Bernard, C. Pancake, and L. Wu. 2006. A Process-Oriented Approach to Engineering Cyberinfrastructure. Report from the Engineering Advisory Committee Subcommittee on Cyberinfrastructure. Available at director.sdsc.edu/pubs/ENG/. Edwards, P., S. Jackson, G. Bowker, C. Knobel. 2007. Understanding InfraÂ�structure: Dynamics, Tensions, and Design. Report of a Workshop on “History & Theory of Infrastructure: Lessons for New Scientific Cyberinfrastructures.” The National Science Foundation. Faniel, I. M. 2009. Unrealized Potential: The Socio-Technical Challenges of a Large Scale Cyberinfrastructure Initiative. NSF Report. NEES (Network on Earthquake Engineering Simulation). 2007. Information Technology within the George E. Brown, Jr. Network for Earthquake Engineering Simulation: A Vision for an Integrated Community. NEES Task Group on Information Technology Vision. Davis, CA. NSF (National Science Foundation). 2007. Cyberinfrastructure Vision for 21st Century Discovery. NSF 07-28. Ribes, D., and T. A. Finholt. 2009. The long now of technology infrastructure: Articulating tensions in development. Journal of the Association for Information Systems 10(5). Sustain. 2009. Position papers from the NSF-sponsored Workshop on Cyberinfrastructure Software Sustainability, March 26-27, 2009, Bloomingdale, IN.
70
A BUILT ENVIRONMENT FROM FOSSIL CARBON John W. Halloran Professor, Department of Materials Science and Engineering University of Michigan Transformative Materials These white papers are intended to stimulate discussion in this Workshop on Grand Challenges in Earthquake Engineering Research. I am a materials engineer with no expertise at all in earthquake engineering, so what can I possibly have to say that could, in some way, stimulate the discussions of the experts at this workshop? This workshop seeks “transformative solutions” for an earthquake-resilient society involving, among other important issues, the design of the physical infrastructure. I will address points relating to the design of the infrastructure, in terms of the most basic materials from which we build our infrastructure. I hope to address transformational change in infrastructural materials that could make our society more resilient not only to the sudden shaking of the ground, but also to the gradual changing of the climate. If you seek a change in the infrastructure large enough to be considered transformational for earthquake resilience, can you also make the change large enough to make a fundamental change in sustainability? I want to attempt to stimulate discussion not only on how to use steel and concrete to make us less vulnerable to shock, but also to make us less vulnerable to climate change. I hope to be provocative to the point of being outrageous, because I want you to think about abandoning steel and concrete. I am going to suggest designing a built environment with more resilient, lighter, stronger, and more sustainable materials based on fossil carbon. The phrase “sustainable materials based on fossil Â�carbon”—seems like an oxymoron. To explain this, I must back up. Fossil resources are obviously not renewable, so are not sustainable in the very long run. But in the shorter run, the limit to sustainability is not the supply of fossil resources but the damage the fossil resources do to our climate. The element of particular concern is fossil-carbon, which was removed from the ancient atmosphere by living creatures and stored as fossil-CO2 in carbonate rocks and as reduced fossilcarbon in hydrocarbons like gas, oil, and coal. The difficulty is that industrial society liberates the fossil carbon to the atmosphere at rates much faster than contemporary photosynthesis can deal with it. It is more sustainable to use fossil resources without liberating fossil-CO2. Can this be done? A Modern World Based on Fossil Life Modern industrial society enjoys prosperity in large part because we are using great storehouses of materials fossilized from ancient life. As the term “fossil fuels” implies, much of our energy economy exploits the fossil residue of
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
photosynthesis. For the energy economy, we all realize that coal and petroleum and natural gas consist of chemically reduced carbon and reduced hydrogen from ancient biomass. Coal and oil are residues of the tissues of green creatures. It is clear that burning fossil-carbon returns to today’s atmosphere the CO2 that was removed by photosynthesis eons ago. We do not often consider that our built environment is also predominantly created from fossils. Cement is made from carbonate limestone, consisting of ancient carbon Â�dioxide chemically combined with calcium oxide in Â�fossil shells. Limestone is one of our planet’s great reservoirs of geo-sequestered carbon dioxide, removed from the air long ago. The reactivity of Portland cements come from the alkaline chemical CaO, which is readily available because limestone fossil rocks are abundant, and the CaO can be obtained by simple heating of CaCO3. However, this liberates about a ton of CO2 per ton of cement. This is fossil-CO2, returned to the atmosphere after being sequestered as a solid. Limestonebased cement—the mainstay of the built Â�environment—is not sustainable. Steel is cheap because we have enormous deposits of iron oxide ore. These iron ores are a kind of geochemical fossil that accumulated during the Great Oxygenation Event 2 billion years ago, as insoluble ferric oxide formed from oxygen produced by photosynthesis. Red iron oxide is a reservoir of the oxygen exhaled by ancient green creatures. We smelt iron ore using carbon from coal, so that carbon fossilized for 300 million years combines with oxygen fossilized for 2,000 million years to flood our current atmosphere with CO2. Every ton of steel is associated with about 1.5 tons of fossil-CO2 liberated in the modern atmosphere. We could, at some cost, capture and sequester the CO2 from limekilns and blast furnaces, or we could choose to smelt iron ore without carbothermic reduction. I prefer to consider a different way to use fossil resources, both in the energy economy and the built environment. We need something besides steel to resist tensile forces, and something besides concrete for resisting compression. In this white paper, I consider using the fossil-hydrogen for energy, and the fossil-carbon for durable materials, an approach called HECAM—Hydrogen Energy and Carbon Materials. This involves simple pyrolysis of coal, oil, or gas to extract fossil-hydrogen for use as a clean fuel. The fossilcarbon is left in the condensed state, as solid carbon or as carbon-rich resins. Production of enough fossil-hydrogen to satisfy the needs of the energy economy creates a very large amount of solid carbon and carbon-rich resins, which can satisfy the needs of the built environment. Fossil Hydrogen for Energy, Fossil Carbon for Materials Fossil fuels and building materials are the substances that our Industrial Society uses in gigatonne quantities. To continue to exploit our fossil resources, and still avoid climate change, we could stop burning fossil-carbon as fuel and
APPENDIX B
stop burning limestone for lime. Coal, petroleum, and gas are used as “hydrogen ores” for energy and “carbon ores” for materials, as presented in more detail previously (Halloran, 2007). Note that this necessarily means that only a fraction of the fossil fuel is available for energy. This fraction ranges from about 60 percent for natural gas to about 20 percent for coal. This might seem like we are “wasting” 40 percent of the gas or 80 percent of the coal—but it is wasted only if a valuable co-product cannot be found. If the solid carbon is used as a building material, the residual carbon could have more value as a solid than it did as a fuel. Natural gas is a rich hydrogen ore, offering about 60 percent of its high heating value (HHV) from hydrogen combustion. It is a relatively lean carbon ore, and if the hydrogen is liberated by simple thermal decomposition: CH4 = 2H2 + C. The solid carbon is deposited from the vapor. Such vapordeposited carbons are usually sooty nanoparticles, such as carbon black, which might be of limited use in the built environment. However, it may be possible to engineer new processes to produce very high strength and highvalue vapor-deposited Â� carbons, such as fibers, nanotubes, of Â�pyrolyic carbons (Halloran, 2008). Pyrolytic carbon has exceptional strength. Vapor-derived carbon fibers could be very strong. Carbon nanotubes, which are very promising, are made from the vapor phase, and the large-scale decomposition of millions of tons of hydrocarbon vapors might provide a path for mass production. An independent path for methane-rich hydrocarbon gases involves dehydrogenating the methane to ethylene: 2CH4 = 2H2 + C2H4. Conversion to ethylene liberates only half the hydrogen for fuel use, reducing the hydrogen energy yield to about 30 percent of the HHV of methane. However, the ethylene is a very useful material feedstock. It can be polymerized to polyethylene, the most important commodity polymer. Polyethylene is mostly carbon (87 wt percent C). Perhaps it is more convenient to sequester the fossil-carbon with some of the fossil-hydrogen as an easy-to-use white resin rather than as more difficult-to-use black elemental carbon. We can consider the polyethylene route as “white HECAM,” with the elemental carbon route as “black HECAM.” Polyethylene from white HECAM could be very useful in the built environment as a thermoplastic resin, the matrix for fiber composites, or a binder for cementing aggregates. It should not be a difficult challenge to produce thermoset grades, to replace hydraulic-setting concretes with chemically setting composites. Moreover, if cost effective methods can be found for producing ultrahigh molecular weight polyethylene (UHMWPE) fibers, we could envision construction materials similar to the current SpectraTM or DyneemaTM fibers, which are among the strongest of all manufactured materials. The tensile strength of these UHMWPE fibers are on the order of 2,400 MPa, more than 10 times higher than the typical yield strength of grade 60 rebar steel. The density of steel is 8 times as large as polyethylene (PE), so the specific strength can be 80 times better for UHWMPE.
71 Petroleum as a hydrogen ore offers about 40 percent of its HHV from hydrogen, and is an exceptionally versatile carbon ore. The petrochemical industry exists to manipulate the C/H ratio of many products producing many structural materials. Indeed carbon fibers—the premier high-strength composite reinforcement—are manufactured from petroleum pitch. Fabricated carbons—the model for masonrylike carbon-building materials—are manufactured from petroleum coke. Coal, with an elemental formulation around CH0.7, is the leanest hydrogen ore (but the richest carbon ore). When coal is heated in the absence of air (pyrolyzed), it partially melts to a viscous liquid (metaplast). Hydrogen and hydrocarbon gases evolve, which swells the viscous metaplast into foam. In metallurgical coke, the metaplast solidifies as a spongy cellular solid. Coke is about as strong as ordinary Portland cement concrete (OPC), but only about one-third the density of OPC. A stronger, denser solid carbon can be obtained by controlling the foaming during pyrolysis by various methods, or by pulverizing the coke and forming a carbon-bondedcarbon with coal tar pitch (a resin from coal pyrolysis). Wiratmoko has demonstrated that these pitch-bonded cokes, similar to conventional carbon masonry, can be 3-10 times as strong at OPC, and stronger that high strength concrete or fired clay brick, at less than half the density of OPC and 60 percent the density of clay brick (Wiratmoko and Â�Halloran, 2009). Like petroleum pitch, coal tar pitch can be used as a precursor for carbon fibers. Although less than 20 percent of the HHV of coal comes from the burning of the hydrogen, coal pyrolysis still can be competitive for the manufacture of hydrogen for fuel. Â�Recently, Guerra conducted a thorough technoÂ�economic analysis of HECAM from coal, using a metallurgical coke plant as a model (Guerra, 2010). Hydrogen could be produced by coal pyrolysis with much less CO2 emission, compared to hydrogen from the standard method of steam reforming of natural gas. The relative hydrogen cost depends on the price of natural gas, the price of coal, and the market value of the solid carbon co-product. Assuming that the solid carbon could be manufactured as a product comparable to concrete masonry blocks, the hydrogen from coal pyrolysis is cheaper if carbon masonry blocks would have about 80 percent of market value of concrete blocks (based on 2007 coal and gas prices). Since the fossil-carbon is not burned in the HECAM process, the carbon dioxide emission is much lower. However, much more coal has to be consumed for the same amount of energy. This carbon, however, is not wasted, but rather put to productive use. Because HECAM combines energy and materials, comparisons of environmental impact are more complicated. For one example (Halloran and Guerra, 2011), we considered producing a certain quantity of energy (1 TJ) and a certain volume of building materials (185 cubic meters). HECAM with hydrogen energy and carbon building materials emitted 47 tons of CO2 and required 236 m3 to
72 be removed by mining. Burning coal and using OPC for the material emitted 150 tons of CO2 and required 245 m3 to be removed from quarries and mines. Can Carbon and Carbon-Rich Solids Be Used in the Infrastructure? The mechanical properties of carbons appear to be Â�favorable. For compression-resistors, carbons have been made that offer significant advantages in strength and strength/density ratio. Carbons are quite inert with respect to aqueous corrosion and should be durable. But of course the properties of these carbons in real environments have never been tested. For tensile-resistors, carbon-fiber composites or UHMWPE-fiber composites should be more than adequate, because they should be stronger and much less dense than steel. Durability has yet to be demonstrated. Fire resistance is an obvious issue. The ability to manufacture these mateÂ� rials in realistic volumes has yet to be demonstrated. An analogue to the setting of cement has not been demonstrated, although conventional chemical cross-linking appears to be viable. Construction methods using these materials have yet to be invented. The cost of HECAM materials is not clearly defined, largely because the materials cost is related to the value of the energy co-product, in the same way that the energy cost is related to the value of the materials co-product. Guerra’s preliminary analysis looks favorable, with each co-product subsidizing the other. Fundamentally, durable materials such as concrete and steel are worth much more per ton than coal, and about the same as natural gas. Values in 2003 were about $70/ton for OPC, $220/ton for rebar, $44/ton for coal, and $88/ton for natural gas (Halloran, 2007). Carbon-rich solids are lower in density (and stronger), so figuring on the basis of volume suggests that converting some of the fossil fuels into construction materials should be economically feasible. But none of this has been demonstrated. Similar carbon materials and composites are known in aerospace technologies as high-performance, very highcost materials. Clearly aerospace-grade graphite fibers or Â�SpectraTM fibers would not possibly be affordable in the tonnages required for infrastructure. For example, the tensile strength of about 1,400 MPa has been reported for carbon Â�fibers produced from coal tar after relatively cheap processing (Halloran, 2007). These are not as strong as aerospacegrade graphite fibers (5,600 MPa), but are comparable in strength to the glass fibers now used in construction, which have a tensile strength of 630 MPa as strands. So we must stretch our imagination to envision construction-grade carbon fibers and UHMWPE fibers, perhaps not as strong but not nearly as costly as aerospace grade. In the same sense, the steel used in rebar is not nearly the cost (or the quality) of the steel used in aircraft landing gear. Much will also depend on when (or if) there will be an effective cost for CO2 emissions. At present in the United
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
States, carbon dioxide from power plants, steel mills, and cement kilns is vented to the atmosphere at no monetary cost to the manufacturer. It is likely in the future that climate change abatement and greenhouse gas control will become a concern for the building industry. Could Carbon-Rich Materials Be Better for Earthquake Resilience? A non-specialist like me, at a workshop of experts like this, should not offer an opinion on this question. I simply do not know. However, two of the strongest and lightest structural materials available for any type of engineering are carbon fibers and UHMWPE fibers, which are fossil-carbon sequestering materials we contemplate for HECAM. The specific strength and specific stiffness of fiber composites based on fossil-carbon based materials should easily exceed the requirements of structural steel. Masonry based on fossilcarbon might easily exceed the performance of ordinary Portland cement concrete, and could be stronger, lighter, and more durable. Would this enable a more earthquake-resilient built infrastructure? Would it make a more environmentally sustainable infrastructure? I hope this is discussed in this workshop. The 19th Century as a Guide for the 21th Century When contemplating any grand challenge, it is useful to look back into history. One hundred years ago, concrete and steel construction was still quite new. One hundred fifty years ago, structural steel was not available for construction. Two hundred years ago, there was no modern cement. So let me go back two centuries, to 1811, and consider what was available for the built environment. There was no structural steel in 1811. Steel was then a very costly engineering material, used in swords and tools in relatively small quantities. Steel was simply not available in the quantity and the cost needed for use as a structural material. There was no Â�Portland Â�cement concrete in 1811. Joseph Aspdin did not patent Â�Portland Â�cement until 1824. But a great deal changed in a rather short time. In 1810, the Tickford Iron Bridge was built with cast iron, not steel. By 1856, Bessemer had reduced the cost of steel, and Â�Siemens had produced ductile steel plate. By 1872, steel was used to build the Brooklyn Bridge. The Wainwright Building in 1890 had the first steel framework. The first glass and steel building (Peter Behrens’ AEG Turbine Factory Berlin) arrived in 1908. I.K. Brunel used Portland cement in his Thames Tunnel in 1828. Joseph Monier used steel-reinforced concrete in 1867, and the first concrete high-rise was built in Cincinnati in 1893. Perhaps a similar change can occur in the 21st century, and perhaps our descendents will think us fools to burn useful materials like carbon.
APPENDIX B
References Guerra, Z. 2010. Technoeconomic Analysis of the Co-Production of Hydrogen Energy and Carbon Materials. Ph.D. Thesis, University of Michigan, Ann Arbor. Halloran, J. W. 2007. Carbon-neutral economy with fossil fuel-based hydrogen energy and carbon materials. Energy Policy 53:4839-4846. Halloran, J. W. 2008. Extraction of hydrogen from fossil resources with production of solid carbon materials. International Journal of Hydrogen Energy 33:2218-2224.
73 Halloran, J. W., and Z. Guerra. 2011. Carbon building materials from coal char: Durable materials for solid carbon sequestration to enable hydrogen production by coal pyrolysis. Pp. 61-71 in Materials Challenges in Alternative and Renewable Energy: Ceramic Transactions, Vol. 224, G. G. Wicks, J. Simon, R. Zidan, E. Lara-Curzio, T. Adams, J. Zayas, A. Karkamkar, R. Sindelar, and B. Garcia-Diaz, eds. John Wiley & Sons, Inc., Hoboken, NJ. Wiratmoko, A., and J. W. Halloran. 2009. Fabricated carbon from �minimally-processed coke and coal tar pitch as a Carbon-Sequestering Construction Material. Journal of Materials Science 34(8):2097-2100.
74
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
UNCERTAINTY QUANTIFICATION AND EXASCALE COMPUTING: OPPORTUNITIES AND CHALLENGES FOR EARTHQUAKE ENGINEERING
equally drawn from a consideration of the �stochastic forward problem, or the stochastic optimization problem.
Omar Ghattas Departments of Mechanical Engineering and Geological Sciences Institute for Computational Engineering & Sciences The University of Texas at Austin
Uncertainty Quantification: Opportunities and Challenges
Introduction In this white paper we consider opportunities to extend large-scale simulation-based seismic hazard and risk analysis from its current reliance on deterministic earthquake simulations to those based on stochastic models. The simulations we have in mind begin with dynamic rupture, proceed through seismic wave propagation in large regions, and ultimately couple to structural response of buildings, bridges, and other critical infrastructure—so-called “ruptureto-rafters” simulations. The deterministic forward problem alone—predicting structural response given rupture, earth, and structural models—requires petascale computing, and is receiving considerable attention (e.g., Cui et al., 2010). The inverse problem—given observations, infer parameters in source, earth, or structural models—increases the computational complexity by several orders of magnitude. Finally, extending the framework to the stochastic setting—where uncertainties in observations and parameters are quantified and propagated to yield uncertainties in predictions—Â�demands the next major prefix in supercomputing: the exascale. Although the anticipated arrival of the age of exascale computing near the end of this decade is expected to provide the raw computing power needed to carry out stochastic Â�rupture-to-rafters simulations, the mere availability of O(1018) flops per second peak performance is insufficient, by itself, to ensure success. There are two overarching challenges: (1) can we overcome the curse of Â�dimensionality to make uncertainty quantification (UQ) for large-scale earthquake simulations tractable, even routine; and (2) can we design efficient parallel algorithms for the deterministic forward simulations at the heart of UQ that are capable of scaling up to the expected million nodes of exascale systems, and that also map well onto the thousand-threaded nodes that will form those systems? These two questions are wide open today; we must begin to address them now if we hope to overcome the challenges of UQ in time for the arrival of the exascale era. We illustrate several of the points in this white paper with examples taken from wave propagation, which is just one component of the end-to-end rupture-to-rafters simulation, but typically the most expensive (some comments are made on the other components). Moreover, we limit the discussion of UQ to the stochastic inverse problem. Despite the narrowing of focus, the conclusions presented here could have been
Perhaps the central challenge facing the field of computational science and engineering today is: how do we quantify uncertainties in the predictions of our large-scale simulations? For many societal grand challenges, the “single point” deterministic predictions issued by most contemporary large-scale simulations of complex systems are just a first step: to be of value for decision making (optimal design, optimal allocation of resources, optimal control, etc.), they must be accompanied by the degree of confidence we have in the predictions. This is particularly true in the field of earthquake engineering, which historically has been a leader in its embrace of stochastic modeling. Indeed, Vision 2025, American Society of Engineers’ (ASCE’s) vision for what it means to be a civil engineer in the world of the future, asserts among other characteristics that civil engineers (must) serve as . . . managers of risk and uncertainty caused by natural events. . . . (ASCE, 2009). Once simulations are endowed with quantified uncertainties, we can formally pose the Â�decision-making problem as an optimization problem governed by stochastic partial differential equations (PDEs) (or other simulation models), with objective and/or constraint functions that take the form of, for example, expectations, and decision variables that represent design or control parameters (e.g., constitutive parameters, initial/boundary conditions, sources, geometry). Uncertainty quantification arises in three fundamental ways in large-scale simulation: • Stochastic inverse problem: Estimation of probability densities for uncertain parameters in large-scale simulations, given noisy observations or measurements. • Stochastic forward problem: Forward propagation of the parameter uncertainties through the simulation to issue stochastic predictions. • Stochastic optimization: Solution of the stochastic optimization problems that make use of statistics of these predictions as objectives and/or constraints. Although solution of stochastic inverse, forward, or optimization problems can be carried out today for smaller models with a handful of uncertain parameters, these tasks are computationally intractable for complex systems characterized by large-scale simulations and high-dimensional parameter spaces using contemporary algorithms (see, e.g., Oden et al., 2011). Moreover, existing methods suffer from the curse of dimensionality: simply throwing more Â�processors at these problems will not address the basic difficulty. We need fundamentally new algorithms for estimation and propagation of, and optimization under, uncertainty in large-scale simulations of complex systems.
75
APPENDIX B
To focus our discussion, in the remainder of this section we will consider challenges and opportunities associated with the first task above, that of solving stochastic inverse problems, and employing Bayesian methods of statistical inference. This will be done in the context of the modeling of seismic wave propagation, which typically constitutes the most expensive component in simulation-based rupture-torafters seismic hazard assessment. The problem of estimating uncertain parameters in a simulation model from observations is fundamentally an inverse problem. The forward problem seeks to predict output observables, such as seismic ground motion at seismometer locations, given the parameters, such as the heterogeneous elastic wave speeds and density throughout a region of interest, by solving the governing equations, such as the elastic (or poroelastic, or poroviscoelastoplastic) wave equations. The forward problem is usually well posed (the solution exists, is unique, and is stable to perturbations in inputs), causal (latertime solutions depend only on earlier time solutions), and local (the forward operator includes derivatives that couple nearby solutions in space and time). The inverse problem, on the other hand, reverses this relationship by seeking to estimate uncertain (and site-specific) parameters from (in situ) measurements or observations. The great challenge of solving inverse problems lies in the fact that they are usually ill-posed, non-causal, and non-local: many different sets of parameter values may be consistent with the data, and the inverse operator couples solution values across space and time. Non-uniqueness in the inverse problem stems in part from the sparsity of data and the uncertainty in both measurements and the model itself, and in part from non-convexity of the parameter-to-observable map (i.e., the solution of the forward problem to yield output observables, given input parameters). The popular approach to obtaining a unique “solution” to the inverse problem is to formulate it as an optimization problem: minimize the misfit between observed and predicted outputs in an appropriate norm while also minimizing a regularization term that penalizes unwanted features of the parameters. This is often called Occam’s approach: find the “simplest” set of parameters that is consistent with the measured data. The inverse problem thus leads to a nonlinear optimization problem that is governed by the forward simulation model. When the forward model takes the form of PDEs (as is the case with the wave propagation models considered here), the result is an optimization problem that is extremely large-scale in the state variables (displacements, stresses or strains, etc.), even when the number of inversion parameters is small. More generally, because of the heterogeneity of the earth, the uncertain parameters are fields, and when discretized result in an inverse problem that is very large scale in the inversion parameters as well. Estimation of parameters using the regularization approach to inverse problems as described above will yield an estimate of the “best” parameter values that simultaneously fit the data and minimize the regularization penalty term.
However, we are interested in not just point estimates of the best-fit parameters, but also a complete statistical description of all the parameter values that is consistent with the data. The Bayesian approach does this by reformulating the inverse problem as a problem in statistical inference, incorporating uncertainties in the measurements, the forward model, and any prior information on the parameters. The solution of this inverse problem is the so-called “posterior” probability densities of the parameters, which reflects the degree of credibility we have in their values (Kaipio and Somersalo, 2005; Tarantola, 2005). Thus we are able to quantify the resulting uncertainty in the model parameters, taking into account uncertainties in the data, model, and prior information. Note that the term “parameter” is used here in the broadest sense—indeed, Bayesian methods have been developed to infer uncertainties in the form of the model as well (so-called structural uncertainties). The Bayesian solution of the inverse problem proceeds as follows. Suppose the relationship between observable outputs y and uncertain input parameters p is denoted by y = f(p, e), where e represents noise due to measurement and/ or modeling errors. In other words, given the parameters p, the function f(p) invokes the solution of the forward problem to yield y, the predictions of the observables. Suppose also that we have the prior probability density πpr(p), which encodes the confidence we have in prior information on the unknown parameters (i.e., independent of information from the present observations), and the likelihood function π(yobs|p), which describes the conditional probability that the parameters p gave rise to the actual measurements yobs. Then Bayes’ theorem of inverse problems expresses the posterior probability density of the parameters, πpost, given the data yobs, as the conditional probability
πpost (p) = π(p|yobs) = k πpr(p) π(yobs|p) (1)
where k is a normalizing constant. The expression (1) provides the statistical solution of the inverse problem as a probability density for the model parameters p. Although it is easy to write down expressions for the posterior probability density such as (1), making use of these expressions poses a challenge, because of the high dimensionality of the posterior probability density (which is a surface of dimension equal to the number of parameters) and because the solution of the forward problem is required at each point on this surface. Straightforward grid-based sampling is out of the question for anything other than a few parameters and inexpensive forward simulations. Special sampling techniques, such as Markov chain Monte Carlo (MCMC) methods, have been developed to generate sample ensembles that typically require many fewer points than gridbased sampling (Kaipio and Somersalo, 2005; Tarantola, 2005). Even so, MCMC approaches will become intractable as the complexity of the forward simulations and the dimension of the parameter spaces increase. When the parameters
76 are a (suitably discretized) field (such as density or elastic wave speeds of a heterogeneous earth), and when the forward PDE requires many hours to solve on a supercomputer for a single point in parameter space (such as seismic wave propagation in large regions), the entire MCMC enterprise collapses dramatically. The central problem in scaling up the standard MCMC methods for large-scale forward simulations and highdimensional parameter spaces is that this approach is purely black-box, i.e., it does not exploit the structure of the Â�parameter-to-observable map f(p). The key to overcoming the curse of dimensionality, we believe, lies in effectively exploiting the structure of this map to implicitly or explicitly reduce the dimension of both the parameter space as well as the state space. The motivation for doing so lies in the fact that the data are often informative about just a fraction of modes of the parameter field, because of ill-posedness of the inverse problem. Another way of saying this is that the Jacobian of the parameter-to-observable map is typically a compact operator, and thus can be represented effectively using a low-rank approximation—that is, it is sparse with respect to some basis (Flath et al., 2011). The remaining dimensions of parameter space that cannot be inferred from the data are typically informed by the prior; however, the prior does not require solution of forward problems, and is thus cheap to compute. Compactness of the parameter-toobservable map suggests that the state space of the forward problem can be reduced as well. A number of current approaches to model reduction for stochastic inverse problems show promise. These range from Gaussian process response surface approximation of the parameter-to-observable map (Kennedy and O’Hagan, 2001), to projection-type forward model reductions (Galbally et al., 2010; Lieberman et al., 2010), to polynomial chaos approximations of the stochastic forward problem (Narayanan and Zabaras, 2004; Ghanem and Doostan, 2006; Marzouk and Naim, 2009), to low-rank approximation of the Hessian of the log-posterior (Flath et al., 2011; Martin et al., In preparation). In the remainder of this section, as just one example of the above ideas, we illustrate the dramatic speedups that can be achieved by exploiting derivative information of the parameter-to-observable map, and in particular the properties of the Hessian. Exploiting derivative information has been the key to overcoming the curse of dimensionality in deterministic inverse and optimization problems (e.g., Akçelik et al., 2006), and we believe it can play a similar critical role in stochastic inverse problems as well. Using modern adjoint techniques, gradients can be computed at a cost of a single linearized forward solve, as can actions of Hessians on vectors. These tools, combined with specialized solvers that exploit the fact that many ill-posed inverse problems have compact data misfit operators, often permit solution of deterministic inverse problems in a dimension-independent (and typically small) number of iterations. Deterministic inverse problems have been solved for millions of parameters and states in tens of
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
iterations, for which the (formally dense, of dimension in the millions) Hessian matrix is never formed, and only its action on a vector (which requires a forward/adjoint pair of solves) is required (Akçelik et al., 2005). These fast deterministic methods can be capitalized upon to accelerate sampling of the posterior density πpost(p), via Langevin dynamics. Long-time trajectories of the Langevin equation sample the posterior, and integrating the equation requires evaluation of the gradient at each sample point. More importantly, the equation can be preconditioned by the inverse of the Hessian, in which case its time discretization is akin to a stochastic Newton method, permitting us to recruit many of the ideas from deterministic large-scale Newton methods developed over the past several decades. This stochastic Newton method has been applied to a nonlinear seismic inversion problem, with the medium parameterized into 65 layers (Martin et al., In preparation). Figure 1 indicates just O(102) samples are necessary to adequately sample the (non-Gaussian) posterior density, while a reference (non-derivative) MCMC method (Delayed Rejection Adaptive Metropolis) is nowhere near converged after even O(105) samples. Moreover, the convergence of the method appears to be independent of the problem dimension when scaling from 65 to 1,000 parameters. Although the forward problem is still quite simple (wave propagation in a 1D layered medium), and the parameter dimension is moderate (up to 1,000 parameters), this prototype example demonstrates the considerable speedups that can be had if problem structure is exploited, as opposed to viewing the simulation as a black box. Exascale Computing: Opportunities and Challenges The advent of the age of petascale computing—and the roadmap for the arrival of exascale computing around 2018—bring unprecedented opportunities to address societal grand challenges in earthquake engineering, and more generally in such fields as biology, climate, energy, manufacturing, materials, and medicine (Oden et al., 2011). But the extraordinary complexity of the next generation of highperformance computing systems—with hundreds of thousands to millions of nodes, each having multiple processors, each with multiple cores, heterogeneous processing units, and deep memory hierarchies—presents tremendous challenges for scientists and engineers seeking to harness their raw power (Keyes, 2011). Two central challenges arise: how do we create parallel algorithms and implementations that (1) scale up to and make effective use of distributed memory systems with O(105-106) nodes and (2) exploit the power of shared memory massively multi-threaded individual nodes? Although the first challenge cited is a classical difficulty, we can at least capitalize on several decades of work on constructing, scaling, analyzing, and applying parallel algorithms for distributed memory high-performance computing systems. Seismic wave propagation, in particular, has had a
77
APPENDIX B
Figure 1╇ Left: Comparison of number of points taken for sampling posterior density for a 65-dimensional seismic inverse problem to identify the distribution of elastic moduli for a layered medium, from reflected waves. DRAM (black), unpreconditioned Langevin (blue), and Stochastic Newton (red) sampling methods are compared. Convergence indicator is multivariate potential scale reduction factor, for which a value of unity indicates convergence. Stochastic Newton requires three orders of magnitude fewer sampling points than the other methods. Right: Comparison of convergence of stochastic Newton method for 65 and 1,000 dimensions suggests dimension independence. SOURCE: Courtesy of James Martin, University of Texas at Austin.
long history of being at the forefront of applications that can exploit massively parallel supercomputing, as illustrated, for example, by recent Gordon Bell Prize finalists and winners (Bao et al., 1996; Akçelik et al., 2003; Komatitsch et al., 2003; Burstedde et al., 2008; Carrington et al., 2008; Cui et al., 2010). To illustrate the strides that have been made and the barriers that remain to be overcome, we provide scalability results for our new seismic wave propagation code. This code solves the coupled acoustic-elastic wave equations in first order (velocity-strain) form using a discontinuous Galerkin spectral element method in space and explicit Runge Kutta in time (Wilcox et al., 2010). The equations are solved in a spherical earth model, with properties given by the Preliminary Reference Earth Model. The seismic source is a double couple point source with a Ricker wavelet in time, with central frequency of 0.28 Hz. Sixth-order spectral elements are used, with at least 10 points per wavelength, resulting in 170 million elements and 525 billion unknowns. Mesh generation is carried out in parallel prior to wave propagation, to ensure that the mesh respects material interfaces and resolves local wavelengths. Table 1 depicts strong scaling of the global seismic wave propagation code to the full size of the Cray XT5 supercomputer (Jaguar) at Oak Ridge National Laboratory (ORNL). The results indicate excellent strong scalability for the overall code (Burstedde et al., 2010). Note that mesh generation costs about 25 time steps (tens of thousands that are typically required), so that the cost of mesh generation is negligible for any realistic simulation. Not only is online parallel mesh generation important for accessing large memory and avoiding large input/output (I/O), but it becomes crucial
for inverse problems, in which the material model changes at each inverse iterations, resulting in a need to remesh repeatedly. The results in Table 1 demonstrate that excellent scalability on the largest contemporary supercomputers can be achieved for the wave propagation solution, even when taking meshing into account, by careful numerical algorithm design and implementation. In this case, a high-order approximation in space (as needed to control dispersion errors) combined with a discontinuous Galerkin formulation (which provides stability and optimal convergence) together provide a higher computation to communication ratio, facilitating
Table 1╇ Strong scaling of discontinuous Galerkin spectral element seismic wave propagation code on the Cray XT-5 at ORNL (Jaguar), for a number of cores ranging from 32K to 224K. # proc cores
meshing time* (s)
wave prop per step (s)
par eff wave
Tflops
32,460 65,280 130,560 223,752
6.32 6.78 17.76 47.64
12.76 6.30 3.12 1.89
1.00 1.01 1.02 0.99
25.6 52.2 105.5 175.6
Meshing time is the time for parallel generation of the mesh (adapted to local wave speed) prior to wave propagation solution; wave prop per step is the runtime in seconds per time step of the wave propagation solve; par eff wave is the parallel efficiency associated with strong scaling; and Tflops is the double precision flop rate in teraflops/s. SOURCE: Courtesy of Carsten Burstedde, Georg Stadler, and Lucas �Wilcox, University of Texas at Austin.
78 better latency tolerance and scalability to O(105) cores, while also resulting in dense local operations that ensure better cache performance. Explicit time integration avoids a global system solve, while the space filling curve-based ordering of the mesh results in better locality. However, if we consider full rupture-to-rafters simulations beyond wave propagation, new and greater challenges arise. Rupture modeling may require dynamic mesh Â�adaptivity to track the evolving rupture front, and historically this has presented significant challenges on highly parallel systems. In recent work, we have designed scalable dynamic mesh refinement and coarsening algorithms that scale to several hundred thousand cores, present little overhead relative to the PDE solution, and support complex gÂ� eometry and high-order continuous/discontinuous discretization (Â�Burstedde et al., 2010). Although they have not yet been applied to dynamic rupture modeling, we expect that the excellent scalability observed in Table 1 will be retained. On the other hand, coupling of wave propagation with structural response presents much greater challenges, because of the need to solve the structural dynamics equations with an implicit method (earthquakes usually excite structures in their low modes, for which explicit methods are highly wasteful). Scalability of implicit solvers to hundreds of thousands of cores and beyond remains extremely challenging because of the global communication required by effective preconditioners, though progress continues to be made (Yang, 2006). Finally, adding nonlinear constitutive models or finite deformations into the soil or structural behavior greatly increases parallel complexity, because of the need for dynamic load balancing and possibly local time stepping. It is fair to say that the difficulties associated with scaling end-to-end rupture-to-rafters simulations are formidable, but not insurmountable if present rates of progress can be sustained (and accelerated). On the other hand, the second challenge identified above—exploiting massive on-chip multithreading—has emerged in the past several years and presents new and pernicious difficulties, particularly for PDE solvers. A sea change is under way in the design of the individual computer chips that power high-end supercomputers (as well as scientific workstations). These chips have exploded in complexity, and now support multiple cores on multiple processors, with deep memory hierarchies and add-on accelerators such as graphic processing units (GPUs). The parallelism within compute nodes has grown remarkably in recent years, from the single core processors of a half-decade ago to the hundreds of cores of modern GPUs and forthcoming many-core Â�processors. These changes have been driven by power and heat dissipation constraints, which have dictated that increased performance cannot come from increasing the speed of individual cores, but rather by increasing the numbers of cores on a chip. As a result, computational scientists and engineers increasingly must contend with high degrees of fine-grained parallelism, even on their laptops and desktop systems, let
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
alone on large clusters and supercomputers. This trend will only continue to accelerate. Current high-end GPUs are capable of a teraflop per second peak performance, which offers a spectacular two orders of magnitude increase over conventional CPUs. The critical question, however, is: can this performance be effectively harnessed by scientists and engineers to accelerate their simulations? The new generation of many-core and accelerated chips performs well on throughput-oriented tasks, such as those supporting computer graphics, video gaming, and high-definition video. Unfortunately, a different picture emerges for scientific and engineering computations. Although certain specialized computations (such as dense matrix problems and those with high degrees of locality) map well onto modern many-core processors and accelerators, the mainstream of conventional scientific and engineering simulations—including the important class of PDE solvers— involve sparse operations, which are memory bandwidthbound, not throughput-bound. As a result, the large increases in the number of cores on a processor, which have occurred without a concomitant increase in memory bandwidth (because of the large cost and low demand from the consumer market), deliver little increase in performance. Indeed, sparse matrix computations often achieve just 1-2 percent of peak performance on modern GPUs (Bell and Garland, 2009). Future peak performance increases will continue to come in the form of processors capable of massive hardware multithreading. It is now up to scientists and engineers to adapt to this new architectural landscape; the results thus far have been decidedly mixed, with some regular problems able to achieve large speedups, but many sparse unstructured problems unable to benefit. Here we provide evidence of the excellent GPU performance that can be obtained by a hybrid parallel CPU-GPU implementation of the discontinuous Galerkin spectral element seismic wave propagation code described above (Â�Burstedde et al., 2010). The mesh generation component remains on the CPU, because of the complex, dynamic data structures involved, while the wave propagation solution has been mapped to the GPU, capitalizing on the local dense blocks that stem from high-order approximation. Table 2 provides weak scaling results on the Longhorn GPU cluster at the Texas Advanced Computing Center (TACC), which consists of 512 NVIDIA FX 5800 GPUs, each with 4GB graphics memory, and 512 Intel Nehalem quad core processors connected by QDR InfiniBand interconnect. The combined mesh generation–wave propagation code is scaled weakly from 8 to 478 CPUs/GPUs, while maintaining between 25K28K seventh-order elements per GPU (the adaptive nature of mesh generation means we cannot precisely guarantee a fixed number of elements). The largest problem has 12.3 million elements and 67 billion unknowns. As can be seen in the table, the scalability of the wave propagation code is exceptional; parallel efficiency remains at 100 percent in weak scaling over the range of GPUs considered. Moreover,
79
APPENDIX B
Table 2╇ Weak scaling of discontinuous Galerkin spectral element seismic wave propagation code on the Longhorn cluster at TACC. #GPUs
#elem
mesh (s)
transf (s)
wave prop
par eff wave
Tflops (s.p.)
8 64 256 478
224048 1778776 6302960 12270656
9.40 9.37 10.6 11.5
13.0 21.3 19.1 16.2
29.95 29.88 30.03 29.89
1.000 1.000 0.997 1.002
0.63 5.07 20.3 37.9
#elem is number of 7th-order spectral elements; mesh is time to generate the mesh on the CPU; tranf is the time to transfer the mesh and other initial data from CPU to GPU memory; wave prop is the normalized runtime (in µsec per time step per average number of elements per GPU); par eff wave is the parallel efficiency of the wave propagation solver in scaling weakly from 8 to 478 GPUs; and Tflops is the sustained single precision flop rate in teraflops/s. The wallclock time of the wave propagation solver is about 1 second per time step; meshing and transfer time are thus completely negligible for realistic simulations. SOURCE: Courtesy of Tim Warburton and Lucas Wilcox.
the wave propagation solver sustains around 80 gigaflops/s (single precision), which is outstanding performance for an irregular, sparse (albeit high-order) PDE solver. Although these results bode well for scaling earthquake simulations to future multi-petaflops systems with massively multi-threaded nodes, we must emphasize that high-orderdiscretized (which enhance local dense operations), explicit (which maintain locality of operations) solvers are in the sweet spot for GPUs. Implicit sparse solvers (as required in structural dynamics) are another story altogether: the sparse matrix-vector multiply alone (which is just the kernel of an iterative linear solver, and much more readily parallelizable than the preconditioner) often sustains only 1-2 percent of peak performance in the most optimized of implementations (Bell and Garland, 2009). Adding nonlinear constitutive behavior and adaptivity for rupture dynamics further complicates matters. In these cases, the challenges of obtaining good performance on GPU and like systems appear overwhelming, and will require a complete rethinking of how we model, discretize, and solve the governing equations. Conclusions The coming age of exascale supercomputing promises to deliver the raw computing power that can facilitate data-driven, inversion-based, high-fidelity, high-resolution rupture-to-rafters simulations that are equipped with quantified uncertainties. This would pave the way to rational simulation-based decision making under uncertainty in such areas as design and retrofit of critical infrastructure in earthquake-prone regions. However, making effective use of that power is a grand challenge of the highest order, owing to the extraordinary complexity of the next generation of computing systems. Scalability of the entire end-to-end process—mesh generation, rupture modeling (including adaptive meshing), seismic wave propagation, coupled structural response, and analysis of the outputs—is questionable on contemporary supercomputers, let alone future exascale systems with three orders of magnitude more cores. Even worse, the sparse, unstructured, implicit, and adaptive nature of rupture-to-
rafters deterministic forward earthquake simulations map poorly to modern consumer-market-driven, throughputoriented chips with their massively multithreaded accelerators. Improvements in computer science techniques (e.g., auto-parallelizing and auto-tuning compilers) are important but insufficient: this problem goes back to the underlying mathematical formulation and algorithms. Finally, even if we could exploit parallelism on modern and emerging systems at all levels, the algorithms at our disposal for UQ suffer from the curse of dimensionality; entirely new algorithms that can scale to large numbers of uncertain parameters and expensive underlying simulations are critically needed. It is imperative that we overcome the challenges of designing algorithms and models for stochastic rupture-torafters simulations with high-dimensional random parameter spaces that can scale on future exascale systems. Failure to do so risks undermining the substantial investments being made by federal agencies to deploy multi-petaflops and exaflops systems. Moreover, the lost opportunities to harness the power of new computing systems will ultimately have consequences many times more severe than mere hardware costs. The future of computational earthquake engineering depends critically on our ability to continue riding the exponentially growing curve of computing power, which is now threatened by architectures that are hostile to the computational models and algorithms that have been favored. Never before has there been as wide a gap between the capabilities of computing systems and our ability to exploit them. Â�Nothing less than a complete rethinking of the entire end-to-end enterprise—beginning with the mathematical formulations of stochastic problems, leading to the manner in which they are approximated numerically, the design of the algorithms that carry out the numerical approximations, and the software that implements these algorithms—is imperative in order that we may exploit the radical changes in architecture with which we are presented, to carry out the stochastic forward and inverse simulations that are essential for rational decision making. This white paper has provided several examples—in the context of forward and inverse seismic wave propagation—of the substantial speedups
80 that can be realized with new formulations, algorithms, and implementations. Significant work lies ahead to extend these and other ideas to the entire spectrum of computations underlying simulation-based seismic hazard and risk assessment. Acknowledgments Figure 1 is the work of James Martin; Table 1 is the work of Carsten Burstedde, Georg Stadler, and Â�Lucas Â�Wilcox; and Table 2 is the work of Tim Warburton and Lucas Â�Wilcox. This work was supported by AFOSR grant FA9550-09-1-0608, NSF grants 0619078, 0724746, 0749334, and 1028889, and DOE grants DEFC02-06ER25782 and DE-FG02-08ER25860. References Akçelik, V., J. Bielak, G. Biros, I. Epanomeritakis, A. Fernandez, O. Â�Ghattas, E. J. Kim, J. Lopez, D. R. O’Hallaron, T. Tu, and J. Urbanic. 2003. High resolution forward and inverse earthquake modeling on terascale computers. In SC03: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE. Akçelik, V., G. Biros, A. Draganescu, O. Ghattas, J. Hill, and B. Van Â�Bloeman Waanders. 2005. Dynamic data-driven inversion for terascale simulations: Real-time identification of airborne contaminants. In Proceedings of the 2005 ACM/IEEE Conference on Supercomputing, Seattle, 2005. Akçelik, V., G. Biros, O. Ghattas, J. Hill, D. Keyes, and B. Van Bloeman Waanders. 2006. Parallel PDE constrained optimization. In Parallel Processing for Scientific Computing, M. Heroux, P. Raghaven, and H. Simon, eds., SIAM. ASCE (American Society for Civil Engineers). 2009. Achieving the Vision for Civil Engineering in 2025: A Roadmap for the Profession. Available at content.asce.org/vision2025/index.html. Bao, H., J. Bielak, O. Ghattas, L. F. Kallivokas, D. R. O’Hallaron, J. R. Shewchuk, and J. Xu. 1996. Earthquake ground motion modeling on parallel computers. In Supercomputing ’96, Pittsburgh, PA, November. Bell, N., and M. Garland. 2009. Implementing sparse matrix-vector multiplication on throughput-oriented processors. In SC09: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE. Burstedde, C., O. Ghattas, M. Gurnis, E. Tan, T. Tu, G. Stadler, L. C. Wilcox, and S. Zhong. 2008. Scalable adaptive mantle convection simulation on petascale supercomputers. In SC08: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE. Burstedde, C., O. Ghattas, M. Gurnis, T. Isaac, G. Stadler, T. Warburton, and L. C. Wilcox. 2010. Extreme-scale AMR. In SC10: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE. Carrington, L., D. Komatitsch, M. Laurenzano, M. M. Tikir, D. Michéa, N. L. Goff, A. Snavely, and J. Tromp. 2008. High-frequency simulations of global seismic wave propagation using SPECFEM3D GLOBE on 62K processors. In SC08: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE.
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH Cui, Y., K. B. Olsen, T. H. Jordan, K. Lee, J. Zhou, P. Small, D. Roten, G. Ely, D.K. Panda, A. Chourasia, J. Levesque, S. M. Day, and P. Â�Maechling. 2010. Scalable earthquake simulation on petascale supercomputers. In SC10: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE. Flath, H. P., L. C. Wilcox, V. Akçelik, J. Hill, B. Van Bloemen Waanders, and O. Ghattas. 2011. Fast algorithms for Bayesian uncertainty quantification in large-scale linear inverse problems based on low-rank partial Hessian approximations. SIAM Journal on Scientific Computing 33(1):407-432. Galbally, D., K. Fidkowski, K. Willcox, and O. Ghattas. 2010. Nonlinear model reduction for uncertainty quantification in large-scale inverse problems. International Journal for Numerical Methods in Engineering 81:1581-1608. Ghanem, R. G., and A. Doostan. 2006. On the construction and analysis of stochastic models: Characterization and propagation of the errors assoÂ� ciated with limited data. Journal of Computational Physics 217:63-81. Kaipio, J., and E. Somersalo. 2005. Statistical and Computational Â�Inverse Problems. Applied Mathematical Sciences, Vol. 160. New York: Springer-Verlag. Kennedy, M. C., and A. O’Hagan. 2001. Bayesian calibration of computer models. Journal of the Royal Statistical Society. Series B (Statistical Methodology) 63:425-464. Keyes, D. E. 2011. Exaflop/s: The why and the how. Comptes Rendus Mcanique 339:70-77. Komatitsch, D., S. Tsuboi, C. Ji, and J. Tromp. 2003. A 14.6 billion degrees of freedom, 5 teraflops, 2.5 terabyte earthquake simulation on the Earth Simulator. In SC03: Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, ACM/IEEE. Lieberman, C., K. Willcox, and O. Ghattas. 2010. Parameter and state model reduction for large-scale statistical inverse problems. SIAM Journal on Scientific Computing 32:2523-2542. Martin, J., L. C. Wilcox, C. Burstedde, and O. Ghattas. A stochastic Newton MCMC method for large scale statistical inverse problems with application to seismic inversion. In preparation. Marzouk, Y. M., and H. N. Najm. 2009. Dimensionality reduction and Â�polynomial chaos acceleration of Bayesian inference in inverse problems. Journal of Computational Physics 228:1862-1902. Narayanan, V. A. B., and N. Zabaras. 2004. Stochastic inverse heat conduction using a spectral approach. International Journal for Numerical Methods Engineering 60:1569-1593. Oden, J. T., O. Ghattas, et al. 2011. Cyber Science and Engineering: A Â�Report of the NSF Advisory Committee for Cyberinfrastructure Task Force on Grand Challenges. Arlington, VA: National Science Foundation. Tarantola, A. 2005. Inverse Problem Theory and Methods for Model Â�Parameter Estimation. Philadelphia, PA: SIAM. Wilcox, L. C., G. Stadler, C. Burstedde, and O. Ghattas. 2010. A highorder discontinuous Galerkin method for wave propagation through coupled elastic-acoustic media. Journal of Computational Physics 229:9373-9396. Yang, U. M. Parallel algebraic multigrid methods—high performance preconditioners. Pp. 209-236 in Numerical Solution of Partial Differential Equations on Parallel Computers, A. Bruaset and A. Tveito, eds., Lecture Notes in Computational Science and Engineering, Vol. 51, Heidelberg: Springer-Verlag.
Appendix C Workshop Participants
Richard Bowker Manager of Information Services Memphis Light, Gas and Water
John Egan Vice President and Principal Engineer AMEC Geomatrix
Mehmet Celebi Senior Research Civil Engineer U.S. Geological Survey
Ronald Eguchi President and CEO ImageCat, Inc.
Adam Crewe Senior Lecturer, Civil Engineering University of Bristol
Ahmed Elgamal Professor, School of Engineering University of California, San Diego
Raymond Daddazio President and CEO Weidlinger Associates, Inc.
Ken Elwood Associate Professor, Department of Civil Engineering University of British Columbia
Gregory Deierlein Professor, School of Engineering Stanford University
Mohammed Ettouney Principal Weidlinger Associates, Inc.
Reginald DesRoches Associate Chair and Professor, School of Civil and Environmental Engineering Georgia Institute of Technology
Charles Farrar Director, Engineering Institute Los Alamos National Laboratory David Feary Senior Program Officer, Board on Earth Sciences and Resources National Research Council
Leonardo Duenas-Osorio Assistant Professor, Department of Civil and Environmental Engineering Rice University
Gregory Fenves Dean, Cockrell School of Engineering University of Texas at Austin
Shirley Dyke Professor, School of Civil Engineering Purdue University
Kent Ferre Geosciences Department Pacific Gas and Electric Company
Eric Edkin Senior Program Assistant, Board on Earth Sciences and Resources National Research Council 81
82 Steve French Associate Dean for Research, College of Architecture Georgia Institute of Technology Omar Ghattas Director, Center for Computational Geosciences University of Texas at Austin
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
Jerome Lynch Director and Associate Professor, Laboratory for Intelligent Structural Technology University of Michigan Stephen Mahin Director, Pacific Earthquake Engineering Research Center University of California, Berkeley
Branko Glisic Assistant Professor, Department of Civil and Environmental Engineering Princeton University
James Malley Vice President, Engineering Degenkolb Engineers
Jerome Hajjar Professor and Chair, Department of Civil and Environmental Engineering Northeastern University
Sami Masri Professor, Department of Civil and Environmental Engineering University of Southern California
John Halloran Professor, Materials Science and Engineering University of Michigan
Peter May Professor, Center for American Politics and Public Policy University of Washington
Thomas Heaton Director, Earthquake Engineering Research Laboratory California Institute of Technology
James Myers Director, Computational Center for Nanotechnology Innovations Rensselaer Polytechnic Institute
Jon Heintz Director of Projects Applied Technology Council William Holmes Principal Rutherford & Chekene Muneo Hori Professor, Earthquake Research Institute University of Tokyo Laurie Johnson President and CEO Laurie Johnson Consulting Kimberly Kurtis Professor, School of Civil & Environmental Engineering Georgia Institute of Technology Kincho Law Professor, School of Engineering Stanford University Bret Lizundia Principal Rutherford & Chekene
Jason Ortego Research Associate, Board on Earth Sciences and Resources National Research Council Shamim Pakzad Assistant Professor, Department of Civil and Environmental Engineering Lehigh University Chris Poland Chairman and CEO Degenkolb Engineers Hope Seligson Associate MMI Engineering Jonathan Stewart Professor and Vice Chair Department of Civil and Environmental Engineering University of California, Los Angeles Costas Synolakis Professor, Department of Civil and Environmental Engineering University of Southern California
83
APPENDIX C
Kathleen Tierney Director, Natural Hazards Research and Applications Information Center University of Colorado at Boulder Sharon Wood Chair and Professor, Department of Civil, Environmental, and Architectural Engineering University of Texas at Austin Solomon Yim Professor, School of Civil and Construction Engineering Oregon State University
OBSERVERS: Joann Culbertson Directorate for Engineering National Science Foundation
Jeffrey Dragovich National Earthquake Hazards Reduction Program National Institute for Standards and Technology Joy Pauschke Directorate for Engineering National Science Foundation Julio Ramirez Director, NEES Operations Center Purdue University Marius Vassiliou Analyst Science and Technology Policy Institute/Institute for Defense Analyses Brian Zuckerman Senior Research Analyst Science and Technology Policy Institute/Institute for Defense Analyses
Appendix D Workshop Agenda
MONDAY, MARCH 14, 2011 8:00 am Breakfast/reception 8:30 Introduction to workshop
Gregory Fenves, Co-chair Chris Poland, Co-chair
8:50
Keynote 1: Community
Laurie Johnson, Laurie Johnson Consulting
9:10
Keynote 2: Lifelines
Reginald DesRoches, Georgia Institute of Technology
9:30
Keynote 3: Buildings
Gregory Deierlein, Stanford University
9:50
Q&A panel session with keynote speakers
10:20 Break 10:35
Briefing on goals of the breakout sessions Instructions given to moderators and participants
10:45 Breakout Session 1 (Dimensional) Define Grand Challenges Pre-event prediction and planning Moderator: John Egan, AMEC Geomatrix Committee rapporteur: Jerome Hajjar, Northeastern University Design of infrastructure Moderator: Ken Elwood, University of British Columbia Committee rapporteur: Adam Crewe, University of Bristol Post-event response and recovery Moderator: Sharon Wood, University of Texas at Austin Committee rapporteur: Jerome Lynch, University of Michigan Community resilience Moderator: Kathleen Tierney, University of Colorado at Boulder Committee rapporteur: Ronald Eguchi, ImageCat, Inc. 12:00 pm Lunch
85
86
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
12:45
Observations from Christchurch
1:00
Plenary Session Workshop participants reconvene to report on breakout session
1:45
Keynote 4: Information Technology Including Q&A
2:15
Breakout Session 2 (Dimensional) Refine Grand Challenges
Ken Elwood, University of British Columbia
James Myers, Rensselaer Polytechnic Institute
3:30 Break 3:45
Breakout Session 3 (Dimensional) Consider facilities
5:00
Plenary Session Workshop participants reconvene to report on breakout session
5:30 Adjourn 6:00
Reception and dinner
TUESDAY, MARCH 15, 2011 8:00 am Breakfast/reception 8:30 Midcourse progress report
Gregory Fenves, Co-chair Chris Poland, Co-chair
9:00
Keynote 5: Materials Including Q&A
John Halloran, University of Michigan
9:30
Keynote 6: Modeling and Simulation Including Q&A
Omar Ghattas, University of Texas at Austin
10:00 Break 10:15 Breakout Session 4 (Disciplinary) Buildings Moderator: Ken Elwood, University of British Columbia Committee rapporteur: Adam Crewe, University of Bristol Lifelines Moderator: Sharon Wood, University of Texas at Austin Committee rapporteur: Jerome Lynch, University of Michigan Geotechnical/Tsunamis Moderator: John Egan, AMEC Geomatrix Committee rapporteur: Jerome Hajjar, Northeastern University Community Moderator: Kathleen Tierney, University of Colorado at Boulder Committee rapporteur: Ronald Eguchi, ImageCat, Inc. 11:50 Lunch Moderators and rapporteurs prepare to discuss results of breakout sessions
87
APPENDIX D
1:00 pm
Observations from Japan
1:20
Plenary Session Workshop participants reconvene to report on disciplinary breakout session
2:50 Final Breakout Session (Dimensional) Reconvene in original breakout groups Pre-event prediction and planning Design of infrastructure Post-event response and recovery Community resilience 4:00
Final Plenary Session
5:00
End Open Session/Adjourn workshop
WEDNESDAY, MARCH 16, 2011 8:00 am
Closed Sessions (open to committee and NRC staff only)
4:00 pm
Adjourn
Thomas Heaton, California Institute of Technology
Appendix E Committee Biographies
Gregory L. Fenves (Co-chair) is Dean of the Cockrell School of Engineering at The University of Texas (UT) at Austin, overseeing the education of more than 7,500 underÂ� graduate and graduate students, and leading some 270 faculty and a research enterprise exceeding $150 million a year. He also is actively involved in improving the commercialization process for intellectual property at UT Austin. Prior to his appointment at UT Austin in September 2008, Dean Fenves served as Chair of the Department of Civil and Environmental Engineering from 2002 to 2007 at the University of California (UC), Berkeley. He served on the faculty of UC Berkeley for more than 20 years and during his tenure there was a key contributor to the Pacific Earthquake Engineering Research Center, a multi-disciplinary engineering program funded by the National Science Foundation. Dean Fenves also served on the research committee for the center’s 10-year duration. In addition, he developed and managed a $12 million industry-sponsored program to improve the seismic safety of utility and transportation systems. An interÂ�nationally recognized structural engineer, Dean Fenves’ research focus is on computational simulation of structures subjected to earthquakes and technology for performancebased engineering. He was one of the early civil engineering researchers to develop wireless sensor networks for assessing the structural health of buildings, bridges, and infrastructure. Dean Fenves earned his bachelor’s degree with distinction from Cornell University and his master’s and doctoral degrees from UC Berkeley. He has received numerous national awards, which include the National Science Foundation’s Presidential Young Investigator Award and the Walter L. Huber Research Prize from the American Society of Civil Engineers.
authority on earthquake engineering, Mr. Poland routinely participates in policy-changing research projects sponsored by the National Science Foundation (NSF), U.S. Geological Survey (USGS), National Institute of Standards and Technology (NIST), and Federal Emergency Management Agency (FEMA). As a passionate advocate and voice for seismic safety, he actively participates in the academic, ethical and social advancement of his field and lectures often. Mr. Poland currently presides as Chair of two con�gressionally mandated advisory committees, on earthquake hazards reduction for the National Earthquake Hazards Reduction Program and on the structural safety of U.S. Department of Veterans Affairs (VA) facilities. His latest interests involving advocacy for resilient cities led to his involvement in the SPUR Resilient City Initiative as the Chair of the Seismic Hazard Mitigation Taskforce. That work led to his Co-chair appointment to the San Francisco Lifelines Council. He chairs the ASCE Standards Committee on Seismic Rehabilitation and led the effort needed to produce the ASCE 31 and ASCE 41 Standards. He is a member of Boards of the San Francisco Chamber of Commerce and the San Francisco Planning and Urban Research Association. He is the 2006 recipient of the Alfred E. Alquist award from the California Earthquake Safety Foundation, and he was recently elected to the National Academy of Engineering in recognition of his career-long work in support of Performance Based Earthquake Engineering. Adam J. Crewe is a senior lecturer in Civil Engineering at the Earthquake Engineering Research Centre at the University of Bristol, United Kingdom. Since joining Bristol University in 1994, he has worked to gain a better understanding of how various types of structure respond to earthquake and other types of dynamic loading, frequently using large-scale laboratory tests to validate his theories. His research areas include civil engineering dynamics, dynamics of bridges, full-scale testing, shaking table modeling, advanced test systems, earthquake resistant design and construction, and public understanding of engineering. He is a member of
Chris D. Poland (NAE) (Co-chair) has a structural engineering career that spans more than 35 years and includes a wide variety of new design work, seismic analysis and strengthening of existing buildings, structural failure analysis, and historic preservation. As an internationally recognized 89
90 many professional bodies, including the Society of Earthquake and Civil Engineering Dynamics, the Institution of Structural Engineers, and the Institution of Civil Engineers. Dr. Crewe received his Ph.D. from Bristol University. Ronald T. Eguchi is President and CEO of ImageCat, Inc., a risk management company specializing in the development and use of advanced technologies for risk assessment and risk reduction. Mr. Eguchi’s areas of expertise include remote sensing, loss estimation, and disaster management, and he has more than 30 years of experience in risk analysis and risk management studies. He has directed major research and application studies in these areas for government agencies and private industry. He currently serves or has served on several editorial boards including the Natural Hazards Review published by the American Society of Civil Engineers and the Natural Hazards Research and Applications Information Center, University of Colorado; the Journal Â� on Uncertainties in Engineering Mechanics published by Resonance Publications, Inc.; and the Earthquake Engineering Research Institute’s (EERI’s) journal SPECTRA. He is currently a member of the National Research Council’s Â�Disaster Roundtable, whose mission it is to identify urgent and important issues related to the understanding and mitigation of natural, technological, and other disasters. He is a past member of the Scientific Advisory Committee of the U.S. Geological Survey, a committee that reports to Congress on recommended research directions for the USGS in the area of earthquake hazard reduction. In 2006, he accepted an ATC Award of Excellence on behalf of the ATC-61 project team for work on “An Independent Study to Assess Future Savings from Mitigation” Activities that showed that a Â�dollar spent on hazard mitigation saves the nation about $4 in future benefits. He was recently recognized by EERI as the 2008 Distinguished Lecturer where he discussed the topic of “Earthquakes, Hurricanes, and Other Disasters: A View from Space.” He was also invited as a keynote speaker to the 14th World Conference on Earthquake Engineering, held in Beijing, China in 2008. He has authored more than 250 publications, many of them dealing with the seismic risk of utility lifeline systems and the use of remote sensing technologies for disaster response. Mr. Eguchi has B.S. and M.S. degrees from the University of California, Los Angeles. Jerome F. Hajjar is professor and Chair at the Department of Civil and Environmental Engineering at Northeastern Â� University, Boston. Previously he was Professor and Â�Narbey Khachaturian Faculty Scholar at the University of Illinois at Urbana-Champaign and Deputy Director of the MidAmerica Earthquake Center. His expertise includes earthquake engineering, computational analysis, experimental
GRAND CHALLENGES IN EARTHQUAKE ENGINEERING RESEARCH
testing, structural stability, and design of steel and composite structures. He serves on the American Institute of Steel Construction Committee on Specifications and is a member of numerous professional societies, including the American Society of Civil Engineers, the Earthquake Engineering Research Institute, and the Network for Earthquake Engineering Simulation Consortium. In 2009 he received the Shortridge Hardesty Award from the American Society of Civil Engineers. Dr. Hajjar received his B.S. in engineering mechanics from Yale University and his M.S. and Ph.D. in structural engineering from Cornell University. Jerome P. Lynch is Director of the Laboratory for Intelligent Structural Technology and Associate Professor in the Department of Civil and Environmental Engineering at the University of Michigan. His research is focused on the development of wireless sensor networks for structural monitoring, the field deployments of wireless monitoring systems, the development of micro- and �nanoelectromechanical system sensors to monitor the response of structures, exploring the piezoresistive properties of engineered high-performance �fiber reinforced cementitious composite, and the development of decentralized algorithms for smart structures. Dr. Lynch received the Office of Naval Research Young Investigator Award in 2005, and the EERI Shah Innovation Award and an NSF CAREER award in 2009. He has a B.E. in civil and environmental engineering from The Cooper Union, M.S. degrees in civil and environmental engineering and in electrical engineering from Stanford University, and a Ph.D. in civil and environmental engineering from Stanford University. Masayoshi Nakashima is Director of the Hyogo Earthquake Engineering Research Center (EDefense), National Research Institute for Earth Science and Disaster Prevention (NIED), and Professor at the Disaster Prevention Research Institute at Kyoto University, Japan. He is also Visiting Professor for the European School for Advanced Studies in Reduction of Seismic Risk at Pavia University, Italy. Dr. Nakashima is Vice President of the Architectural Institute of Japan, where he also chairs the Research Committee on Structures, and he serves on the board of directors at the Earthquake Engineering Research Institute (EERI). He was council member for the Science Council of Japan from 2000 to 2003. In April 2009, he received the Special Achievement Award from the American Institute of Steel Construction for sustained strong support of collaboration between the United States and �Japan on research to improve the seismic performance of steel building structures. Dr. Nakashima received his B.S.C.E. and M.S.C.E. from Kyoto University and his Ph.D. from Lehigh University.