Back to Table of contents

Primeur weekly 2018-07-03

Quantum computing

Scientists pump up chances for quantum computing ...

CEA uses Atos simulator at the CCRT to explore the potential of quantum computing in industry ...

Focus on Europe

European parliament votes in favour of EuroHPC JU ...

European Commission welcomes Parliament vote on plans to establish EuroHPC JU ...

Report European Investment Bank: Financing the future of supercomputing - How to increase investments in high performance computing in Europe ...

Looking back on a successful PRACEdays18 ...

Middleware

Aerodynamic science reveals the best position in a peloton of 121 cyclists and calculates unexpected drag reduction for the athletes thanks to the largest simulation ever done in sport ...

Hardware

WatermelonBlock Partners with IBM Watson using its Supercomputer to let crypto investors know what the market is thinking ...

Tech Automotive Leaders Join Forces on Next-Generation In-Vehicle Networking Technologies for Autonomous and Connected Vehicles ...

Chayora Celebrates Major Milestone as It Nears Completion of Phase 1 of Its First Facility at Its Tianjin Hyperscale Data Centre Campus in China ...

Indian Institute of Technology Bombay deploys Cray to Power Research and Education ...

Lisa Compute Cluster expanded with GPUs for machine learning ...

Mellanox announces agreement with Starboard ...

Applications

Simulation for grid transmission and distribution ...

Futuristic data storage ...

GA4GH streaming API htsget a bridge to the future for modern genomic data processing ...

Department of Energy taps Argonne to lead effort focused on energy-water systems ...

Researchers discover new enzyme paradigm for critical reaction in converting lignin to useful produce useful products ...

SenseTime debuts in Singapore by signing MOU with local giants NTU, NSCC and Singtel ...

New simulations break down potential impact of a major quake by building location and size ...

AI learns the art of debate ...

Tiny sensors may help avert earthquake damage, track sonar danger, "listen" to pipelines ...

BNAs improve performance of Li-ion batteries ...

RUDN chemists have completely changed the direction of Diels-Alder reaction ...

UMass Amherst geoscientists offer new evidence for how the Adirondack Mountains formed ...

A galactic test will clarify the existence of dark matter ...

The Cloud

HPE accelerates data insight and action across the enterprise with latest Edgeline capabilities ...

Hewlett Packard Enterprise commits $4 billion to accelerate the Intelligent Edge ...

New simulations break down potential impact of a major quake by building location and size


This image resulting from recent simulations at NERSC show the distribution of ground motion intensity across the San Francisco Bay Area region 10 seconds after a large-magnitude earthquake along the Hayward Fault. Credit: Berkeley Lab
28 Jun 2018 Berkeley - With unprecedented resolution, scientists and engineers are simulating precisely how a large-magnitude earthquake along the Hayward Fault would affect different locations and buildings across the San Francisco Bay Area.

A team from Lawrence Berkeley National Laboratory (Berkeley Lab) and Lawrence Livermore National Laboratory, both U.S. Department of Energy (DOE) national labs, is leveraging powerful supercomputers to portray the impact of high-frequency ground motion on thousands of representative different-sized buildings spread out across the California region.

Their work - part of DOE's Exascale Computing Project - is important to showing how different seismic wave frequencies of ground motion affect structures of various sizes. Lower-frequency ground motion is known to affect larger structures and is easier to replicate by computer simulation. Small structures like homes are more vulnerable to high-frequency shaking, which requires more advanced computing to simulate.

The researchers are presenting three scientific papers describing their recent simulations at this week's U.S. National Conference on Earthquake Engineering (NCEE), a meeting held every four years by the Earthquake Engineering Research Institute. One simulation was run last week using the Cori supercomputer at Berkeley Lab's National Energy Research Scientific Center (NERSC) to simulate high-frequency (5-hertz) ground shaking.

David McCallen, a senior scientist in the Earth and Environmental Sciences Area at Berkeley Lab, explains that their simulations allow scientists to gain a more realistic picture of the impact that a major earthquake would have on a region.

"Historically, seismic experts have used empirical evidence from previous earthquakes to assess earthquake hazard and risk at regional scale," said McCallen. "While important, data about ground motion characteristics and resulting structural damage extrapolated from an earthquake that occurred halfway across the world are only so helpful to improving our understanding of how well our California infrastructure could endure seismic stress."

At the NCEE conference taking place in Los Angeles this week, the team will describe recent simulation advancements indicating that event magnitude and ground displacement are positively correlated; that two buildings with the same number of stories 2.4 miles apart and equidistant to the fault line could suffer damages to a far different degree; and that three-story buildings are less sensitive than 40-story buildings to the significant increase in long-period ground motion (exceeding 1 second) that would accompany a large-magnitude earthquake.

At Berkeley Lab, McCallen leads a program focused on integrating advanced technologies and data analysis to enable risk-informed seismic design of critical infrastructure, such as bridges and the power grid. He says that the ability to produce such high-resolution, physics-based simulations as the team has done on the Cori system represents a transformational era now underway for assessing earthquake hazard (ground motion) and the resulting risk (building damage) on a regional scale.

Because it runs through the East Bay, the most populated subregion of the San Francisco Bay Area, the Hayward Fault is considered one of the most dangerous faults in the United States. The fault hasn't generated a major earthquake since 1868, a source of concern for experts who cite evidence supporting the idea that the East Bay fault is overdue for a major earthquake.

The national lab researchers leveraged Cori's supercomputing power to simulate ground motions at a broad range of frequencies, which are then used in a second computer program for building response, to quantify seismic risk for representative building structures on a regional scale. A total of 9,600 structural dynamic simulations were employed and analyzed to study the risk variation on a 100-kilometer by 50-kilometer domain for two events: one at magnitude 6.5, and another at magnitude 7.

A critical factor affecting earthquake damage to buildings and structures is seismic wave frequency, or the rate at which an earthquake wave repeats each second. Because of this, researchers including Livermore Lab's computational scientist Anders Petersson and seismologist Arthur Rodgers have been working with Berkeley Lab's Hans Johansen to advance the existing SW4 code. This code was originally developed by Petersson for simulating three-dimensional seismic wave propagation.

"While working closely with the NERSC operations team in a simulation last week, we used essentially the entire Cori machine - 8,192 nodes, and 524,288 cores - to execute an unprecedented 5-hertz run of the entire San Francisco Bay Area region for a magnitude 7 Hayward Fault earthquake. The run was executed in 9 hours and 11 minutes of wall clock time," said McCallen. "The code developments necessary for this run are moving us toward our ultimate goal of a full exascale application."

The fact that buildings respond differently to certain seismic wave frequencies based on their size was evidenced by the most recent simulations of high-magnitude earthquakes along the Hayward Fault. These showed an increase in damage potential for the 40-story building - more than for the three-story building - as the earthquake increased in magnitude from 6.5 to 7 at 5 hertz due to the significant increase in ground motion at longer periods of vibration. Other results indicate that ground motion and the potential for structural damage could vary between sites in relatively close proximity.

Their simulations showed that two buildings with the same number of stories equidistant from the fault line and only approximately 3 miles apart can have a substantially different damage potential due to the differences in which seismic waves emanating from the fault merge together. Buildings in proximity to the fault rupture and fault-rupture directivity zone were found to be at greater risk of damage from the earthquakes, and fault-slip-associated permanent ground displacement was shown to increase with event magnitude.

Without the NERSC computing resources making it possible to execute these earthquake event scenarios for so many locations and structures at such high resolution so quickly, it would not be possible to produce such accurate site-specific depictions of structural risk. The researchers believe that such precise physics-based representations of the potential for damages will ultimately provide a more accurate and complete picture of the interaction between seismic ground motion and structural damage.

"Simulations can both increase understanding of and reduce uncertainties about the very complex processes at play in earthquake science and engineering," said McCallen. "It's become clear that our ability to accurately assess earthquake hazard and risk at regional scale can benefit from physics-based simulations that allow us to depict the impact of seismicity on site-specific ground motion and structures in the areas most vulnerable to these potentially devastating events."

"It is an exciting time for the science and engineering communities. Through DOE's Exascale Project, we will be developing the computational ecosystem and gaining access to computers that are big enough and fast enough to execute these types of computationally intensive calculations," he added.

Source: DOE/Lawrence Berkeley NationalLaboratory

Back to Table of contents

Primeur weekly 2018-07-03

Quantum computing

Scientists pump up chances for quantum computing ...

CEA uses Atos simulator at the CCRT to explore the potential of quantum computing in industry ...

Focus on Europe

European parliament votes in favour of EuroHPC JU ...

European Commission welcomes Parliament vote on plans to establish EuroHPC JU ...

Report European Investment Bank: Financing the future of supercomputing - How to increase investments in high performance computing in Europe ...

Looking back on a successful PRACEdays18 ...

Middleware

Aerodynamic science reveals the best position in a peloton of 121 cyclists and calculates unexpected drag reduction for the athletes thanks to the largest simulation ever done in sport ...

Hardware

WatermelonBlock Partners with IBM Watson using its Supercomputer to let crypto investors know what the market is thinking ...

Tech Automotive Leaders Join Forces on Next-Generation In-Vehicle Networking Technologies for Autonomous and Connected Vehicles ...

Chayora Celebrates Major Milestone as It Nears Completion of Phase 1 of Its First Facility at Its Tianjin Hyperscale Data Centre Campus in China ...

Indian Institute of Technology Bombay deploys Cray to Power Research and Education ...

Lisa Compute Cluster expanded with GPUs for machine learning ...

Mellanox announces agreement with Starboard ...

Applications

Simulation for grid transmission and distribution ...

Futuristic data storage ...

GA4GH streaming API htsget a bridge to the future for modern genomic data processing ...

Department of Energy taps Argonne to lead effort focused on energy-water systems ...

Researchers discover new enzyme paradigm for critical reaction in converting lignin to useful produce useful products ...

SenseTime debuts in Singapore by signing MOU with local giants NTU, NSCC and Singtel ...

New simulations break down potential impact of a major quake by building location and size ...

AI learns the art of debate ...

Tiny sensors may help avert earthquake damage, track sonar danger, "listen" to pipelines ...

BNAs improve performance of Li-ion batteries ...

RUDN chemists have completely changed the direction of Diels-Alder reaction ...

UMass Amherst geoscientists offer new evidence for how the Adirondack Mountains formed ...

A galactic test will clarify the existence of dark matter ...

The Cloud

HPE accelerates data insight and action across the enterprise with latest Edgeline capabilities ...

Hewlett Packard Enterprise commits $4 billion to accelerate the Intelligent Edge ...