The number of ground motions used in nonlinear response history analysis (NRHA) determines the precision of the parameter estimates obtained in seismic performance assessments. While this issue has been extensively studied in the earthquake engineering literature, the relationship of probability model misspecification to parameter estimation uncertainty, and the implication to the required number of ground motions needed for NRHA, has not been examined. Probability model misspecification has the potential to increase estimation uncertainty and hence requires a greater number of ground motions to achieve the same precision compared to when misspecification is disregarded. This study develops a procedure to determine the required number of ground motions in seismic code-prescriptive and risk-based assessments with possible probability model misspecification. Specifically, we employ the quasi-maximum likelihood estimation (QMLE), which is robust to probability model misspecification, to evaluate estimation uncertainty. The QMLE approach is applied to an archetype California bridge under the two seismic assessment scenarios. In the code-prescriptive assessment, misspecification errors are identified for dispersion estimates of the bridge column ductility demand. In the most extreme case of the risk-based evaluation, misspecification increases the estimation uncertainty of the mean annual frequency of exceeding a limit state by as much as three times, which substantially increases the required number of ground motions. Based on the findings from this study, we advocate for the use of QMLE to detect and rectify the implications of model misspecification to estimation uncertainty and the number of ground motions used in probabilistic seismic performance assessments.
Maximum likelihood estimation (MLE) is commonly used in seismic performance assessment procedures. Parameters in a probability model are estimated based on data either from structural analysis or post-earthquake reconnaissance by maximizing a likelihood function. MLE-based fitting is used in different modules within the performance-based earthquake engineering framework including in the development of ground motion models (GMMs; Boore et al., 2014), seismic demand and capacity models (Cornell et al., 2002), fragility models (Baker, 2015; Lallemant et al., 2015), and loss functions (Aslani, 2005). The lognormal distribution is the most commonly specified functional form in earthquake engineering because of its relationship to the normal distribution, and its performance in goodness-of-fit tests relative to other probability models (e.g. Ibarra, 2004).
MLE-based fragility fitting procedures contain an implicit assumption, that is the probability model (or the corresponding likelihood function) is “correctly specified,” meaning the specified model is aligned with the underlying data-generating process. For example, Stafford et al. (2022) used the inverse of the Fisher information (IF) matrix to develop ground motion logic trees. Using the IF method to derive standard errors implies that the model is assumed to be correctly specified (White, 1982). The probabilistic demand model developed in Cornell et al. (2002) and many follow-up studies adopted the homoscedasticity assumption, meaning the error term is assumed to be constant across all ground motion intensity levels. This assumption is prone to misspecification errors. Similarly, Lallemant and Kiremidjian (2017) and Sáez et al. (2011) studied the standard errors in fragility parameters assuming the correct probability model specification.
Prior studies have shown that the correct specification assumption may lead to underestimation of the estimation uncertainty (e.g. Angrist and Pischke, 2009: Chapter 9; King and Roberts, 2015). To identify and rectify misspecification, two classes of approaches can be utilized: model class selection and standard error amendment. Model class selection uses the best-fit model from a set of candidates based on evaluation metrics like Akaike and Bayesian information criteria. Muto and Beck (2008) used Bayesian updating and model class selection to identify the best-suited hysteretic shear-building models among a class of candidate hysteretic models. Jalayer et al. (2023) used the Bayesian model class selection method to determine the best fragility function among a suite of candidate models. On the other hand, one can choose to amend the standard error of the parameters embedded in the misspecified probability model without directly altering it. White (1982) proposed the quasi-MLE (QMLE) method where the asymptotic distribution of the parameter estimates was tailored to a misspecified model. QMLE yields asymptotically consistent standard errors (meaning as the sample size approaches infinity, the probability that the standard error estimates deviate from their true values is zero), named the robust standard error, even if the model is not correctly specified. Dahal et al. (2022) compared the estimation uncertainty in building collapse risk between the classical MLE and QMLE methods. However, the observation that the standard errors from the QMLE method are consistently smaller than the conventional standard errors may not always hold, which is discussed in this study. Also, the Dahal et al. (2022) study did not consider the implications of model misspecification to the size of the record-set needed for structural response analysis.
In earthquake engineering, the standard error of the inferred model parameters is important because it reveals the degree of estimation uncertainty. Theoretically, the estimation uncertainty is zero if an infinite number of ground motions are used. However, for practical purposes, response history analysis is performed with a finite number of ground motions, . In general, as is increased, there will be a reduction in the estimation uncertainty. Several studies evaluated the relationship between and estimation uncertainty (e.g. Baltzopoulos et al., 2019; Sáez et al., 2011). However, these studies did not consider model misspecification. As such, the derived was prone to underestimation.
The required number of ground motions is the size of the record-set needed to limit the estimation uncertainty in the parameters of interest such that the desired level of precision is achieved. The precision of the parameters can be quantified by their corresponding inferential statistics, including the coefficient of variation , sample mean and standard deviation, and confidence interval. Bradley (2011) used the 84th percentile of the sample mean seismic demand as a design criterion. Buratti et al. (2011) used confidence intervals of the median and dispersion to quantify the precision of engineering demand parameters (EDPs). Gehl et al. (2015) specified of the median and logarithmic standard deviation that define a seismic demand model as precision measures. Jalayer et al. (2017) integrated Bayesian inference and cloud analysis to obtain an updated fragility function with a limited number of ground motions by marginalizing the posterior distribution of the associated parameters. The confidence intervals of fragility functions were used as a measure of their precision. Kiani et al. (2018) and Baltzopoulos et al. (2019), respectively, used the of mean estimates of the annual rate of exceeding specified EDP values and the of the collapse risk to determine the required number of ground motions. It is notable that none of the aforementioned studies considered probability model misspecification, which could lead to biased standard errors and overestimation of the precision in the associated parameters.
The standard error of a statistical estimate is a measure of the uncertainty in the parameter and reliability of the inference. Standard errors can be obtained using analytical or simulation-based methods. The analytical method is built on the asymptotic theories including the law of large numbers and central limit theorem, and is derived based on the Fisher information contained in the inferred model parameters (Lallemant and Kiremidjian, 2017; Sáez et al., 2011; Stafford et al., 2022). Simulation-based methods use the bootstrap technique (Efron, 2000) in which the sample space is treated as if it were the population, and realizations are repeatably generated with replacement. The bootstrap standard error is obtained from the distribution constructed by implementing the resampling scheme. Baltzopoulos et al. (2019) and Baker (2015) used the simulation-based approach to obtain standard errors for fragility parameters. In this study, we focus on the analytical method.
The used in structural analyses has implications to both code-prescriptive (e.g. in the structural design phase where a single (or multiple) intensity level is of interest) and risk-based assessments (a range of intensity levels are considered and the mean annual frequency of exceedance is of interest). The minimum number of ground motions is typically specified in seismic design codes or guidelines. For example, Caltrans Seismic Design Criterion v2.0 (SDC v2.0; California Department of Transportation, 2018) specifies that a minimum of 7 ground motions should be used for response history analysis of bridges. Whereas ASCE 7-22 (American Society of Civil Engineers, 2022) requires a minimum of 11 ground motions, FEMA P-58 (Federal Emergency Management Agency, 2012) specifies a minimum of 7 ground motions for seismic response and loss assessments. The minimum specified in these codes, standards, and guidelines assume that reliable estimates can be obtained from the resulting sampling distribution. However, no substantive reasoning is provided for the choice of the minimum . In this study, we benchmark the required minimum specified by Caltrans against both code-based and risk-based bridge performance assessments while explicitly considering model misspecification.
This study addresses the issue of misspecification as it relates to code- and risk-based assessments and the used in structural response analysis. Given that the lognormal distribution is the most commonly used model in earthquake engineering, we propose to maintain this practice but use the QMLE method to enhance the robustness of the inference. The rest of the article proceeds as follows. The next section presents the necessary theoretical background for both the classical IF-based and robust approaches to computing standard errors. A simulation-based study is conducted to demonstrate the benefits of the robust approach for two hypothetical model misspecification scenarios in seismic analysis procedures. Next, the information matrix test, which detects the presence of modal misspecification through hypothesis testing, is introduced. The robust standard error approach is applied to determine the implications to the used to assess the performance of typical California bridges. Three levels of seismicity are considered and for each site, three design eras are adopted (i.e. the bridges have varying capacities). Using the of the parameter estimates as the measure of precision, the relationship to the required used in code-based and risk-based assessments is investigated.
The classical and robust standard errors
This section describes the formulation of the robust standard error. We begin by using to denote a parametric family, where is a -dimensional vector (for lognormal distribution, ). Given that there are observations of seismic response demands that are drawn from the underlying probability density function , denotes the true parameter that defines the probability model. The maximum likelihood estimator, , is obtained by maximizing the log-likelihood function .
The classical standard error (i.e. the IF-based standard error ), which is derived from the central limit theory with the assumption that the probability model is correctly specified, is:
where is the inverse of the information matrix obtained from observations evaluated at the maximum likelihood estimator , and denotes the diagonal elements of a matrix. The observed information matrix serves as an alternative to the theoretical Fisher information matrix, , which is typically infeasible to calculate in advance. The proof of Equation 1, along with additional details, is provided in Supplement A of the supplemental material for this article.
Equation 1 is built on the assumption that the true data-generating distribution is within the specified parametric family, . However, in reality, the parametric family may not contain the functional form of . This is the case when model misspecification occurs. For example, the lognormal distribution family which is typically specified in seismic demand, capacity, and collapse fragility models may not be consistent with the inherently true model that generated the seismic data.
The QMLE method (White, 1982) aims to yield an asymptotically consistent standard error for the ML estimator, even if the likelihood function is incorrect. The main idea of QMLE is, without knowing the specific form of , we aim to find an alternative within the parameter space that minimizes the distance between and . The robust standard error for the ML estimator based on QMLE theory has the following form:
where and are, respectively, the plug-in versions of the bread matrix and meat matrix evaluated at . is defined as the negative expectation with respect to of the Hessian matrix, while is defined as the variance with respect to of the score function. This standard error estimator is robust since it makes no assumptions that the probability model is correctly specified. The step-by-step formulation is provided in Supplement A.
Note that the theoretical IF-based standard error should always be smaller than the theoretical robust standard error. This is because the variance of the IF-based ML estimator achieves the Cramér–Rao Lower Bound, which means among all the unbiased estimators for , the IF-based variance is the lowest. However, when using real data, the robust standard errors are sometimes lower than the IF-based version (e.g. Dahal et al., 2022). This is because (1) in practice they are calculated from their plug-in versions (i.e. Equations 1 and 2, respectively), and (2) in the context of the Cramèr–Rao Lower Bound, the IF-based variance achieves this lower bound asymptotically, but in an application context, the sample size is limited. Having a robust standard error that is less than the IF-based counterpart is also a sign that the model misspecification is mild. A related simulation-based study by Angrist and Pischke (2009: Chapter 9) concluded that in cases with significant misspecification, the mean of is always greater than the mean of , but the former exhibits greater sampling variability. In contrast, the relation between the mean of and is uncertain in cases with little or no misspecification. They attributed the situation where the robust standard error is smaller than the IF-based standard error to sample bias and the inflated sampling variability in the robust standard error. Therefore, we adopt the recommendation of Angrist and Pischke (2009: Chapter 9) and Imbens and Kolesar (2016) to use the maximum of the IF-based standard error and robust standard error to truncate lower values of robust estimators and reduce sample bias:
To provide some intuition regarding the differences between the two distinct standard errors, Figure 1 depicts the estimated standard errors for the two parameters in a lognormally specified fragility curve—the median of intensity measure , , and dispersion (or, log-standard deviation), , based on structural response data from previously published articles (Dastmalchi and Burton, 2021; Haindl et al., 2023). The data contain counts of building structures exceeding different limit states subjected to earthquake ground motion excitations at different intensity levels. The observation that the robust standard errors deviate from their IF-based counterparts signifies that the specification of the lognormal form is incorrect. This means that a greater number of ground motions is needed to achieve the precision that is obtained from the IF-based method. Later we introduce a method to detect misspecification through hypothesis testing.
for (a) and (b) calculated from two types of standard errors based on data from the literature.
Simulation-based investigation
To demonstrate the superiority of the robust standard error over the classical standard error and reason intuitively about their differences, we create two hypothetical model misspecification scenarios that may arise in seismic assessments: misspecifying the link function and violating the independence assumption.
Generalized linear models (GLMs), which are often used for fragility fitting, are comprised of two components: the response variable , which is a member of the exponential family, and a link function that relates the conditional expectation of the response variable to the linear combination of predictors. In the case of a limit state fragility, the response variable is a dichotomous number, and only one predictor (i.e. ) exists. Therefore, the GLM has the form:
where denotes the probability of attaining or exceeding some predefined limit state under a specified level.
If the probit link function is used (i.e. where is the standard normal cumulative density function (CDF)), and the predictor is replaced by its logarithmic form, , then fitting the fragility based on the GLM is mathematically equivalent to using a lognormal CDF (e.g. Baker, 2015):
The relationship between fragility parameters and GLM coefficients is as follows: is the median, and is the dispersion. With these relationships, the uncertainty in the GLM coefficients and can be transformed to that of the lognormal fragility parameters and .
Scenario 1: Misspecified link function
This subsection discusses the case where the specified probit link in a GLM deviates from the true link function that generates the data. Procedures for this simulation-based investigation are summarized below:
Specify the true GLM coefficients and in Equation 4.
Specify a true link function (other than the probit link) from which the probability of limit state exceedance is generated: .
for (where is the total number of simulations) do
(a) Generate random binomial numbers indicating whether or not there is limit state exceedance.
(b) Use the misspecified probit link function to fit the GLM, and obtain and .
(c) Estimate the standard errors of and using the two distinct methods, which are denoted as and , respectively.
Now we have a set of -simulated GLM coefficients, . The sample standard deviation from the simulated GLM coefficients is denoted as the empirical true standard error, .
Construct 95% confidence intervals from the sampling distributions of and .
The goal is to compare the 95% confidence intervals from the two types of standard errors and the empirical true value to demonstrate the superiority of the robust standard error. A total of simulations are performed. A fixed number of hypothetical ground motions is used at each IM level. The levels are defined from 0.1 to 4 with an increment of 0.1 (the unit is not important here).
Table 1 shows the result of this simulation study. The 95% confidence intervals from the robust standard errors (the last column) produce wider intervals than that of the IF-based method (the fifth column), and the former always contains the empirical true standard errors (the fourth column). In contrast, the confidence intervals from the IF-based method fail to capture the empirical true values. This indicates that the robust method is able to accurately estimate standard errors even when the probability model is misspecified.
Results from the simulation-based investigation
True GLM coefficient
Empirical true
95% confidence interval from the IF method
95% confidence interval from the robust method
−2
2
20
cloglog
0.117
0.128
−2
2
40
cloglog
0.084
0.092
−2
2
100
cloglog
0.054
0.060
−2
4
20
cauchit
0.117
0.146
−2
4
40
cauchit
0.081
0.102
−2
4
40
logit
0.083
0.106
−2
4
100
logit
0.055
0.072
The robust standard error manifests its robustness when the link function is misspecified. GLM = generalized linear model.
Scenario 2: Violation of the independent observation assumption
An important assumption that is embedded in binomial random variables is that the individual trial observations are independent. This is also the assumption adopted in seismic fragility fitting procedures, that is, each observation of structural failure or not is assumed to be independent of all the others. However, this assumption may not be valid in many cases. For example, ground motions that are selected from the same earthquake event may cause dependency in seismic responses. Also, research on ground motion rotation considers different incidence angles for a fixed record, which violates the independence assumption.
In this simulation setting, we draw dependent binomial numbers using the beta-binomial distribution (Wilcox, 1981), in which the probability of structural failure is dependent on a specific beta distribution:
which means the probability of observing failures out of ground motions is dependent on the beta distribution parameters and . is the beta function. A tuning parameter is used to control the degree of dependence. We use levels from 0.1 to 4 with an increment of 0.1. The true probability of failure at the level is assumed to be . A total of 500 simulations are performed. The probit link is specified as the probability model. The results in Table 2 show similar trends as the preceding stimulation study. The confidence intervals from the robust approach always contain the empirical true standard errors, while the confidence intervals from the IF-based method do not. This demonstrates the superiority of in estimating the standard errors of GLM coefficients when the independence assumption is violated.
Results from the simulation-based investigation
Empirical true
95% confidence interval from the IF method
95% confidence interval from the robust method
0.15
0.146
0.171
0.15
0.135
0.160
0.15
0.134
0.154
0.3
0.192
0.213
0.3
0.184
0.204
0.5
0.247
0.274
0.5
0.250
0.276
The robust standard error manifests its robustness when the independence assumption is violated
The information matrix test
Several methods exist for diagnosing probability model specification. For instance, the quantile–quantile plot allows for visual comparisons of the data against the presumed distribution (e.g. Jayaram and Baker, 2008). On the other hand, hypothesis testing can be employed if the data conforms to the presumed distribution.
In this article, we use the information matrix test (a variant of hypothesis testing) initially introduced by White (1982) to detect misspecification. By comparing Equations 1 and 2, misspecification exists if the meat matrix (the middle matrix in the robust standard error) diverges from the observed information matrix. This leads to the null and alternative hypotheses:
The information matrix test identifies whether there exists a statistically significant divergence between the two matrices. The method proposed in Dhaene and Hoorelbeke (2004) is adopted, which used a nested parametric bootstrap to calculate the -value of the information matrix test.
To derive the test statistic, first, for and obtained from which are binary numbers corresponding to the exceedance of a certain limit state or not, we calculate . The matrix is then stacked into a vector of length : . The test statistic is expressed as , where is the consistent estimate of under . To estimate and the subsequent -value, we adopt the nested bootstrap in Dhaene and Hoorelbeke (2004), which is summarized below:
The information matrix test
1: Create an empty list with a length of .2: fordo3: Sample , where is the ML estimate of the fragility function.4: Fit a GLM based on , from which we calculate and stack it into .5: fordo6: Repeat rows 3 and 4 and yield for 7: end for8: Calculate , where 9: Calculate 10: Set .11: end for12: Calculate , where .13: Calculate .14: -value = for .
Figure 2 shows the -values of the information matrix test for the same data used in Figure 1. By specifying a significance level of 0.05, the is rejected in six out of ten cases which implies misspecification.
p-Values from the information matrix test based on data from the literature.
Description of bridges, structural modeling, and performance limit states
The effects of misspecification are investigated with respect to probabilistic seismic responses and risk analyses of California bridges. The goal is to uncover how the adoption of the robust standard error influences the estimation uncertainty in the seismic response and risk.
Description of bridges and nonlinear structural modeling
Two-span box girders are one of the most commonly used bridge superstructures in the state of California. As such, two-span seat-type bridges with a single column designed in different eras are used as the case study structures.
The bridge nonlinear structural modeling strategy in Mangalathu (2017) is adopted. Three-dimensional bridge models are constructed and analyzed using OpenSeesPy (Zhu et al., 2018).
The bridge superstructure comprises elasticBeamColumn elements with concentrated masses along the centerline, assuming the superstructure maintains elasticity during an earthquake. The transverse connection between the bridge deck and the abutment is modeled using a rigidLink element. The column bents are modeled using dispBeamColumn elements with four integration points. The sections are discretized into steel (Steel02), and unconfined and confined concrete (Concrete07) fibers.
Translational and rotational linear springs are added at the column base to represent the foundation–soil interaction. The response of the abutment in the longitudinal direction is governed by passive and active resistance. Passive resistance is developed in parallel by the backfill soil and abutment piles when the abutment backwall moves toward the backfill soil. Active resistance is developed when the abutment moves away from the backfill soil and is influenced by the abutment piles only. The abutment transverse response is resisted by the piles only. The backfill soil is modeled using zero-length elements in the longitudinal direction, with the HyperbolicGap material. The pile behaviors in the longitudinal and transverse directions are modeled using bi-directional zero-length elements, with the hysteretic material.
The expansion joint between the deck and abutment consists of three components: the elastomeric bearings, pounding between the deck and abutment, and shear keys. The bi-directional behavior of the bearings is simulated via the Steel01 material. The shear key in the transverse direction is modeled using the trilinear material with a gap. The pounding in the longitudinal direction is modeled using the ElasticPPGap material.
Model parameter uncertainty is not considered. Therefore, the associated parameters are taken as their expected (for normal distributed variables) or median values (for lognormal distributed variables) based on the probability models defined in Mangalathu (2017). A mass factor, serving as a multiplier applied to the original mass, accounts for the fact that superstructure components, such as parapets and rails, are not explicitly considered in the bridge model. The same geometry (including the span length, deck width, column height, etc.) is used for bridges designed in different eras. The bridge modeling parameters are summarized in Table 3. Readers can refer to Figure 3.2 in Mangalathu (2017) for the bridge model schematic.
Model parameters for the seat-type bridges
Name
Unit
Value
Name
Unit
Value
Span length
mm
31,775
Abutment pile lateral capacity per deck width
N/mm
1498
Deck width
mm
9780
Elastomeric bearing stiffness per deck width
N/mm/m
908
Column height
mm
6625
Elastomeric bearing frictional coefficient
–
0.3
Column diameter
mm
60
Shear key acceleration
g
1
Abutment height
mm
2186
Gap—longitudinal
mm
23.5
Unconfined concrete compressive strength
MPa
29.03
Gap—transverse
mm
12.8
Reinforcement steel yield strength
MPa
465.0
Mass factor
–
1.25
Column pier—translational stiffness
N/mm
297,716
Damping ratio
%
4.5
Column pier—rotational stiffness
N-m/rad
4.5E+9
Longitudinal reinforcement ratio
%
2.25
Bridge performance limit states
Three bridge design eras are considered, including pre-1971 (Era1), 1971–1990 (Era2), and post-1990 (Era3). The differences among the design eras are reflected in the component and system seismic capacities. Bridge components are categorized into primary or secondary. Primary components include the columns and abutment seats, which affect bridge safety. Secondary components include the abutment piles and backfill soil, expansion joints, and decks, whose failure affects serviceability but does not compromise safety. The bridge is treated as a multicomponent system, where system failure is determined by the failure of its ten constituent components. A series system is assumed where the exceedance counts for the bridge are calculated as the envelope of the component exceedance counts (Li et al., 2020). The component-level limit states for bridges designed in three eras are shown in Table 4 (Mangalathu, 2017).
Component-level limit states
Limit states
Component
EDP
Unit
S
M
E
C
Primary components
Column
Era1
Curvature ductility
–
0.8
2.0
5.0
8.0
Era2
1.0
5.0
8.0
11.0
Era3
1.0
5.0
11.0
17.0
Abutment seat
Era1Era2Era3
Superstructure unseating
inch
0.51.01.5
1.04.54.5
2.010.014.0
3.015.021.0
Secondary components
Bent foundation—translational
Translational foundation response
inch
1.0
4.0
NA
NA
Bent Foundation—rotational
Rotational foundation response
rad
1.5
6.0
NA
NA
Abutment passive
Passive abutment response
inch
3.0
10.0
NA
NA
Abutment active
Active abutment response
inch
1.5
4.0
NA
NA
Abutment transverse
Transverse abutment response
inch
1.0
4.0
NA
NA
Deck
Deck displacement
inch
4.0
12.0
NA
NA
Bearing
Bearing displacement
inch
1.0
4.0
NA
NA
Shear key
Shear key displacement
inch
1.0
5.0
NA
NA
S = slight damage; M = moderate damage; E = extensive damage; C = collapse; NA = not applicable; EDP = engineering demand parameter.
Hazard characterization
Probabilistic seismic hazard analysis (PSHA) is performed at the sites of interest. Varying levels of seismic hazard are considered by selecting three sites in California located in Oakland (high seismicity), Malibu West (moderate seismicity), and Sacramento (low seismicity). The time averaged shear-wave velocity for the upper 30-m depth is taken as , assuming all the sites are in NEHRP soil class BC (Building Seismic Safety Council, 2003; i.e. ). The seismicity-level descriptors for these three sites are defined relative to other cities in California. The site information is shown in Table 5. The three coordinates are based on the US Geological Survey (USGS) reference locations (Petersen et al., 2015).
Site information
Site
Seismic level
Latitude
Longitude
Oakland
High
37.8
−122.25
760
Malibu West
Moderate
34.05
−118.95
Sacramento
Low
38.6
−121.5
The geometric mean spectral acceleration over a range of periods, (where is the spectral acceleration at period and is the number of structural periods chosen) is utilized as the . This is chosen because a bridge is typically dominated by more than one mode of vibration. has been shown to be a more efficient for probabilistic seismic demand and performance assessment for structures affected by higher mode response (Cordova et al., 2000; Eads et al., 2013; Vamvatsikos and Cornell, 2005). The first six model periods are as follows: (longitudinal), (transverse), (torsion), (longitudinal), (longitudinal), and (transverse). is computed as the geometric mean from to with a uniform period spacing of .
PSHA is performed using the OpenQuake Hazard Characterization engine (Pagani et al., 2014). We use the UCERF3 model (Field et al., 2014) to characterize rupture rates for the seismic sources in California. The GMM proposed by Boore et al. (2014) is specified to predict the seismic intensity in active shallow crustal tectonic zones. The constituent in is developed based on 5%-damped orientation-independent spectral coordinate, (Boore, 2010), where the directionality of the two orthogonal horizontal ground motion pairs is considered. The subsequent ground motion selection and scaling schemes are also based on , but we omit the subscript for simplicity. The correlation model for at different periods developed by Baker and Jayaram (2008) is used to quantify the dependence for the calculation.
Six levels with return periods of 50, 225, 475, 975, 2475, and 10,000 years are considered in the risk-based assessment. The 225- and 975-year return periods correspond to, respectively, the functional and safety-evaluation earthquakes in the Caltrans SDC v2.0 (California Department of Transportation, 2018). The 475- and 2475-year return periods are defined as the design-basis and maximum-considered earthquakes, respectively, in ASCE 7-22 (American Society of Civil Engineers, 2022).
The hazard curves and conditional mean spectra (CMSs; Baker, 2011) for the three sites are shown in Figure 3. The CMS is approximated based on the results of mean hazard disaggregation. It is apparent that Oakland has the highest seismicity, while Sacramento has the lowest.
(a) Hazard curves and (b) at a return period of 975 years, for the three sites.
Quantifying the effects of model misspecification on the required number of ground motions
Seismic analysis procedures for parametric probability models involve: (1) selecting a finite number of ground motions that represent the site hazard, (2) performing nonlinear response history analyses (NRHAs) based on the selected ground motions, (3) fitting a probability model based on the structural response data, and (4) making statistical inferences about the parameters that define the probability model. Due to the finite sample size, using different sets of ground motions yields different parameter estimates. The variability in the parameter estimates is described as the estimation uncertainty.
This section investigates the effects of the number of ground motions on parameter estimation uncertainty, both in code-prescriptive and risk-based assessments. Misspecification potential and its impact on estimation uncertainty and the required number of ground motions are also explicitly considered. Different inferential statistics are used for the two seismic evaluation procedures. The code-based assessment focuses on the precision of the descriptive statistics (e.g. mean, median, and standard deviation) for the EDPs of interest, while the risk-based assessment considers the precision of the mean annual frequency of exceeding different performance limit states. Specifically, the of the median and dispersion estimates of a critical bridge EDP are considered for the code-based assessment. In the risk-based assessment, we use the of the mean annual frequency of exceeding the slight, moderate, extensive, and collapse limit states.
Since most prior research and applications related to probabilistic seismic performance assessments specify the lognormal distribution as the functional form, we stick to using this probability model in both the code-prescriptive and risk-based procedures.
Code-prescriptive assessment
A code-prescriptive assessment in which NRHA is performed at a single or multiple levels is commonly adopted in seismic design procedures (Baker et al., 2021: Chapter 10). The goal of this type of assessment is to estimate the parameters that define the conditional (on the ) probability distributions of the critical EDPs and ensure that the summary statistic (e.g. the mean) is below some predefined threshold. A minimum number of ground motions that must be selected based on a target spectrum is typically specified in seismic codes, standards, and guidelines. This minimum number is specified under the assumption that it will yield reliable estimates from the resulting sample.
The model for a specific EDP that expresses the probability of exceeding certain values conditional on an is given by:
where and are the conditional median and dispersion for the EDP of interest. We are interested in the uncertainty in the maximum likelihood estimates for these two parameters because they control the required number of ground motions.
The relationship between the estimation uncertainty of the maximum curvature ductility of bridge columns and the number of ground motions is established. The Caltrans SDC v2.0 design spectrum is chosen as the target. The same document specifies the uniform hazard spectrum (UHS) at a return period of 975 years as the design spectrum. It is based on the 2014 version of the USGS seismic hazard map (Petersen et al., 2015), and adjusts for the UHS by incorporating amplification factors to account for near-fault and basin effects.
We choose as the EDP of interest because it is often used as a measure of bridge safety and overall performance (e.g. Kowalsky et al., 1995; Zahn et al., 1986). is the ratio between the maximum absolute curvature and the column yield curvature, . The latter is obtained from a column section analysis as .
To investigate the impact of the number of ground motions on the estimation uncertainty of , we vary from 3 to 200. For each , we calculate the of and . One strategy is to repeatably select ground motions that are consistent with the design spectrum, perform NRHA, and estimate the inferential statistics, from . Recall the Caltrans design spectrum is a single mean spectrum, such that the candidate ground motions are selected based on a ranking of their individual mean-squared errors (MSEs). Therefore, we use the following approach: first, 200 ground motions are selected from the PEER NGA-West2 database (Bozorgnia et al., 2014) to match the design spectrum. Then, the records with the lowest MSE in the 200 ground motion set are chosen. The median and dispersion estimates for as well as their estimation uncertainty are obtained based on the subset. Figure 4 shows the Caltrans design spectra, as well as the mean spectra of the selected 200 ground motions, for each of the three sites.
Selected ground motions that match the Caltrans Seismic Design Criterion v2.0 design spectrum.
Figure 5 shows the estimates from the IF-based and robust methods for and for different sizes of record-sets used in the NRHAs. For comparison purposes, we specify two variability levels: a moderate variability level corresponding to a CV of 0.3, and a low variability level with a CV of 0.15. This setting is subjective and can be adjusted depending on specific engineering needs. As expected, the estimates in general decrease as increases; however, there are some fluctuations, especially when a small sample size is used. This is because the standard errors are derived in the asymptotic context. Although they are inversely proportional to , their plug-in estimates do not guarantee a monotonically decreasing trend.
(a) and (b) versus the used in the design phase.
The estimates for decrease sharply from to , and converge gradually after , for all three sites. In general, there is no difference between the results from the IF-based and robust methods (notice that they overlap in Figure 5a). This means the specification of the lognormal functional form for the median estimation is essentially correct. Also, observe that the site with the highest seismicity (Oakland) exhibits the highest , while the site with the lowest seismic activity (Sacramento) shows the lowest.
The estimates for (Figure 5b) without considering model misspecification are the same for all three sites. Supplement B shows that, for any normally or lognormally distributed random variables, the estimate of dispersion is independent of specific sample values. This means that if the goal is to determine the minimum based on the estimation uncertainty of without considering the possibility of misspecification, this can be done without performing any upfront structural analysis. This finding is exclusive to .
There is obvious evidence of misspecification when using the lognormal form to estimate the dispersion. For example, is significantly larger than when for Oakland. Misspecification is more severe for Sacramento, which happens when . For Malibu West, model misspecification is milder compared to Oakland and Sacramento. However, there is evidence showing slight misspecification when . The occurrence of model misspecification in dispersion estimation suggests that caution should be taken when using the analytical form in Supplement B, which is based on the correct specification assumption. The from the IF-based method is 22 if the goal is to have low variability in for the Sacramento site. This number increases to 41 when the robust standard error is considered. Therefore, QMLE is recommended when the precision of the dispersion is of interest.
The Caltrans SDC v2.0 Section 4.2.3 requires at least 7 ground motions for NRHAs. In our example application, produces median estimates with a low variability level . This indicates that using 7 ground motions is sufficient to yield reliable estimates for the median EDP of the considered bridges. In addition, misspecification is identified in dispersion estimates at all sites across various ranges of record-set sizes, which is evidenced by the discrepancy between the two types of standard errors. This happens in the range of 8–16 for Oakland, 25–200 for Sacramento, and 91–200 for Malibu West. This is a sign of misspecification since deviance should not happen if the model is correctly specified.
Does the increased matter in engineering practice?
As shown in Figure 5b, divergences between and are observed in some ranges of . The most extreme case is in Sacramento when a set of ground motions is used, which has , , and .
The increased dispersion only matters if it has substantial impacts on the overall variability of the bridge responses. To further investigate this, we perform a simulation study based on the median and dispersion estimates in the most extreme case. For each realization , the following procedure is performed:
Based on the statistical moments of and under , generate new samples: , , and .
Generate two sets of responses with using different standard deviation estimates. One is drawn by . The other is drawn by . The difference is that the first set is sampled using the standard deviation drawn from the IF-based standard error, while the second set uses the robust one.
Calculate the sample mean and sample standard deviation from the two sets.
We repeat the above procedure 1000 times and compare the empirical distributions of the sample mean and standard deviation of from the two sets. The results show the sample means of the simulated from the two methods are the same (1.23), and the sample standard deviations are also equal (0.72). Thus, the increased estimation uncertainty in the dispersion does not impact the variability of the bridge column ductility responses.
Risk-based assessment
In this section, we investigate the effects of the used at each level in the risk-based assessment on the precision of mean annual frequency of exceedance estimates. Multiple stripe analysis (MSA) is used for the bridge seismic performance assessment. At each level (stripe), records are selected based on the corresponding target conditional spectrum (CS). With a CS-based target, MSA has been shown to be an efficient (Baker, 2015) and hazard-consistent (Chandramohan et al., 2016; Jalayer and Cornell, 2009) structural analysis strategy. We choose to set the CS as the target because we do not want to introduce additional uncertainty due to hazard inconsistency. Note that the CS only matches the IM of interest and other parameters such as duration are not considered. This limitation was overcome by using the generalized CS developed in Bradley (2010).
A discrete set of is used. Unlike the Caltrans design spectrum, CS contains two target spectra that are matched simultaneously. In this case, the strategy of ground motion selection used in the code-prescriptive assessment does not work, because a good match with the CMS does not guarantee a good match with the conditional standard deviation spectrum. Instead, we perform distinct ground motion selections at each level and each site. A total of NRHAs are performed for each value. EzGM (Özsaraç, 2023) is used to automate the ground motion selection procedure. For illustration purposes, Figure 6 shows that the selected record-set matches the target CS for the 975-year return period at the Oakland site, with .
Selection of 20 ground motions based on the CS for the Oakland site and 975-year return period. (a) Conditional mean matching and (b) conditional dispersion matching.
The mean annual frequency of exceeding a given limit state, , is calculated by integrating the fragility function with the hazard curve:
The uncertainty in the hazard curve is not considered, that is, the mean hazard curve from PSHA is used. In this sense, the differences in the seismic hazard are constant, and the uncertainty only comes from the fragility term.
It is difficult to derive the moments of analytically since the probability terms involve a function of random variables. Instead, Lallemant and Kiremidjian (2017) used the first-order second-moment approximation to estimate the mean and variance of the mean annual frequency:
where is the standard deviation of , and is the correlation coefficient between and . Since and are expressed with the same GLM coefficients and , they are perfectly correlated and hence .
Figure 7 shows the estimates for the MAF of exceeding different limit states (expressed as ) for for the slight, moderate, extensive, and collapse limit states, with used in each stripe, for the three design eras and three sites. Some of the estimates are not reported and are described as “uninformative” cases because almost all ground motions cause the bridge to exceed (or not) the associated limit state. The uninformative cases issue is further discussed later in this section.
for different limit states and values used at each level.
When model misspecification is ignored, exhibits a decreasing trend as increases. Similar to the code-prescriptive assessment, this decreasing trend is not monotonic. When approaches 30, tends to converge. However, it is obvious that model misspecification is prevalent. Figure 8 reports the -values from the information matrix test introduced before. Although in some cases the -values from the hypothesis testing are relatively high (-value ) and the is therefore not rejected, we still observe significant differences between the estimates from the two methods. For example, for the Era 1 bridge at the Malibu West site using 40 ground motions is nearly twice that of its IF counterpart. However, the evidence is not strong enough to reject (its -value = 0.27). This is because the selection of an appropriate significance level is subjective, and the choice of the significance level in this study may be unconservative.
-Values from the information matrix test for different limit states versus used at each level.
The increase in the in the misspecification cases indicates a greater number of ground motions are required to achieve a target precision. For example, when ignoring the possibility of misspecification, using 5 ground motions at each level for the Era 1 bridge located at the Sacramento site is sufficient to lower the estimation uncertainty to the moderate variability level (with ). However, model misspecification increases this value to , which is 30% higher than the IF-based method. This indicates that at least 10 ground motions should be used to achieve the same precision. Similarly, for the Era 2 bridge located at the Malibu West site using 10 ground motions has two times the estimation uncertainty compared to its IF-based counterpart. In this situation, twice the number of ground motions is needed to limit to the moderate variability level.
The estimation uncertainty across different sites does not contain any clear trend. However, sites with low (Sacramento) and moderate (Malibu West) seismicity generally show higher estimation uncertainty than the site with the highest seismicity (Oakland) for the Era 1 and Era 2 bridges. In addition, in many cases, the for slight and moderate limit states is higher than the more severe limit states (extensive and collapse). For the Era 1 and Era 2 bridges, 5 ground motions per stripe basically achieves a moderate variability level for the extensive and collapse limit states, and 30 ground motions are needed to reduce the estimation uncertainty to the low variability level. For slight and moderate damage, this number increases to 20 for moderate variability, and roughly 60 to achieve low variability. Also, there is no significant difference observed for across different bridge design eras. This is particularly true for the IF-based method.
Discussion of the uninformative cases. is not reported when uninformative observations are present. This happens when a bridge has very low capacity (e.g. slight limit state for the Era 1 design) and is subjected to high seismic demands (Oakland), or in bridges with very high capacity (e.g. collapse capacity of Era 3 bridges) located at a site with low seismic hazard (Sacramento). In these cases, the limit state exceedance counts are uninformative, meaning under all of the levels, nearly all ground motions cause the bridge to fail or survive. When this happens, the algorithm determines the probability of exceeding a limit state to be nearly deterministic with infinite estimation uncertainty. Thus, both the IF-based and the robust standard errors fail.
To better visualize the difference between the uninformative and informative cases, Figure 9 shows the log-likelihood surfaces as a function of the two GLM coefficients. The red points marked at the surface are the maximum likelihood estimates. Recall the observed information matrix or the bread matrix (shown in Equations A.14 and A.21 of the supplement, the negative second derivative of the log-likelihood function) essentially represents the curvature of the surface around the maximum likelihood estimator. A small curvature (the left panel) indicates low information carried by the data about the true parameters, and thereby high estimation uncertainty. Conversely, a high curvature (the right panel) implies low estimation uncertainty.
Comparison of log-likelihood surfaces between (a) an uninformative case (Malibu West, Era 2, collapse) and (b) an informative case (Malibu West, Era1, extensive damage).
We propose several solutions to these uninformative observations:
One can simply claim the structure is strong enough to resist any level of shaking up to the maximum considered level, or is incapable of resisting any shaking above the minimum considered level, according to the current structural design scheme and limit state. However, this claim is highly uncertain since there exists no more information above or below the levels under consideration.
Incorporating even lower or higher levels into the MSA. This strategy may lead to effective exceedance counts (i.e. non-zero or non-one proportions) at extremely low or high levels. However, except for special structures (e.g. nuclear plants), in most cases, it is not necessary to consider extremely rare or common events. In this study, the selected levels with return periods ranging from 50 to 10,000 years are appropriate.
Increasing the number of ground motions used in each stripe. The uninformative case may be due to the use of a very small sample size, which can easily lead to selection bias.
Using an alternative structural analysis strategy. For example, cloud analysis (Jalayer et al., 2017), in which a greater number of unscaled ground motions are selected to match a single target spectrum at a fixed rate of exceedance. Since the ground motions are unscaled, the levels are continuous, and it is possible to obtain more effective counts than MSA. However, as suggested by Jalayer et al. (2017), a very careful ground motion selection should be performed to ensure a great portion of records leads to structural failures and the dispersion in the records’ intensity is considerable. We chose to use MSA coupled with CS because it is recognized to be a more efficient (Baker, 2015) and hazard-consistent (Chandramohan et al., 2016) strategy than the cloud analysis.
Conclusion
Estimation uncertainty is inevitable in seismic performance assessment procedures due to the use of a finite record-set size. The required number of ground motions for NRHAs is evaluated based on the precision of the statistical estimates of interest. To measure the precision of parameter estimates, inferential statistics that are based on standard errors of the parameter must be estimated from samples. This study introduces two types of standard errors which are derived based on the asymptotic normality of maximum likelihood estimators. One is based on the inverse of the Fisher information matrix, which does not consider misspecification potential. The other is the robust standard error that stems from QMLE theory which maintains asymptotic consistency even if the probability model is misspecified.
We develop a procedure that quantifies the estimation uncertainty in both code-prescriptive and risk-based performance assessments of California bridges. In the code-prescriptive assessment, we select varying numbers of ground motions that are consistent with the Caltrans SDC v2.0 design spectra. Choosing the CV of the median and dispersion as target inferential statistics, we investigate the effects of on the estimation uncertainty in the column maximum curvature ductility . Specific observations are as follows:
In general, misspecification is not detected in . Using either the IF-based or robust standard errors yields the same . However, misspecification is detected for dispersion estimates. The most severe case happens in a low seismic hazard zone, which increases the estimation uncertainty by a factor of 1.9 and leads to nearly twice the number of ground motions being needed to achieve a target precision.
For , a higher estimation uncertainty is observed in locations with high seismicity. For , we found that the estimates of any normally or lognormally distributed EDP are independent of specific sample values if misspecification potential is ignored. This means that analysts can immediately determine from the derived analytical formula in Supplement B without performing any upfront structural analyses. However, since misspecification is detected in dispersion estimates, we suggest using QMLE when the primary interest is in the precision of dispersion.
The specification of a minimum of 7 ground motions in Caltrans SDC v2.0 for bridge design is evaluated. This minimum requirement is essentially sufficient to control the precision of the median EDP estimate at the low variability level and the estimation uncertainty for dispersion has moderate variability. If the goal is to have a more precise estimate of the dispersion, at least 20 ground motions should be selected.
In the risk-based assessment, we use MSA based on different sets of CS-consistent ground motions while varying the used in each stripe. Bridges with varying seismic capacities that were designed in different eras, and three sites in California with high, moderate, and low seismic hazard levels, are considered.
We choose the CV of the mean annual frequency of limit state exceedance as the target inferential statistic. Specific observations are as follows:
Overlooking misspecification yields biased estimation uncertainty and hence an underestimated required number of ground motions. In contrast, using the robust standard error to amend estimation uncertainty when misspecification is present leads to a greater required record-set size. This is revealed by the fact that the cases detected as having misspecified probability models make the monotonic decreasing trend of estimation uncertainty as a function of sample size unstable.
Generally, for the slight and moderate limit states has higher uncertainty than for the more severe limit states. For the Era 1 and Era 2 bridges, roughly 5 and 30 ground motions are needed for the extensive damage and collapse limit states to attain the moderate and low variability levels, respectively. These numbers are increased to roughly 20 and 60 for the slight and moderate limit states, respectively. Also, sites with low and moderate seismicity generally show higher estimation uncertainty than the high seismicity site.
The uninformative observations in the risk-based assessment are discussed, which occur when the limit state is exceeded (or not) for all or nearly all the ground motions and levels. When there are uninformative observations, the corresponding statistical inference is invalid. We propose several strategies to address this issue.
The procedure developed in this study can be used to detect model misspecification, quantify estimation uncertainty in seismic performance assessment procedures, and determine the number of ground motions that are required to yield a reliable estimate of the parameters of interest. We recommend analysts use QMLE to detect model misspecification and rectify the estimation uncertainty when misspecification is present. While the findings and conclusions from this study are specific to the considered bridges, the proposed framework can be implemented for other types of structures.
Supplemental Material
sj-pdf-1-eqs-10.1177_87552930241262044 – Supplemental material for Effects of probability model misspecification on the number of ground motions required for seismic performance assessment
Supplemental material, sj-pdf-1-eqs-10.1177_87552930241262044 for Effects of probability model misspecification on the number of ground motions required for seismic performance assessment by Chenhao Wu and Henry V Burton in Earthquake Spectra
Footnotes
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: The research is supported by the Pacific Earthquake Engineering Research Center’s Transportation Systems Research Program Topic 2: Forward Uncertainty Quantification. Any opinions, findings, and conclusions expressed in this article are those of the authors and do not necessarily reflect the views of the sponsor.
Data availability
The R codes that were used for analyses have been made available at .
Supplemental material
Supplemental material for this article is available online.
References
1.
American Society of Civil Engineers (2022) Minimum design loads for buildings and other structures. ASCE/SEI 7-22. Reston, VA: ASCE.
2.
AngristJDPischkeJ-S (2009) Mostly Harmless Econometrics: An Empiricist’s Companion. Princeton, NJ: Princeton University Press.
3.
AslaniH (2005) Probabilistic Earthquake Loss Estimation and Loss Disaggregation in Buildings. PhD thesis, Stanford University, Stanford, CA.
4.
BakerJBradleyBStaffordP (2021) Seismic Hazard and Risk Analysis. Cambridge: Cambridge University Press.
5.
BakerJW (2011) Conditional mean spectrum: Tool for ground-motion selection. Journal of Structural Engineering137(3): 322–331.
6.
BakerJW (2015) Efficient analytical fragility function fitting using dynamic structural analysis. Earthquake Spectra31(1): 579–599.
7.
BakerJWJayaramN (2008) Correlation of spectral acceleration values from NGA ground motion models. Earthquake Spectra24(1): 299–317.
8.
BaltzopoulosGBaraschinoRIervolinoI (2019) On the number of records for structural risk estimation in PBEE. Earthquake Engineering & Structural Dynamics48(5): 489–506.
9.
BooreDM (2010) Orientation-independent, nongeometric-mean measures of seismic intensity from two horizontal components of motion. Bulletin of the Seismological Society of America100(4): 1830–1835.
10.
BooreDMStewartJPSeyhanEAtkinsonGM (2014) NGA-West2 equations for predicting PGA, PGV, and 5% damped PSA for shallow crustal earthquakes. Earthquake Spectra30(3): 1057–1085.
11.
BozorgniaYAbrahamsonNAAtikLAAnchetaTDAtkinsonGMBakerJWBaltayABooreDMCampbellKWChiouBS-JDarraghRBDaySDonahueJGravesRWGregorNHanksTCIdrissIMKamaiRKishidaTKottkeAMahinSARezaeianSRowshandelBSeyhanEShahiSShantzTSilvaWSpudichPAStewartJPWatson-LampreyJWooddellKYoungsR (2014) NGA-West2 research project. Earthquake Spectra30(3): 973–987.
12.
BradleyBA (2010) A generalized conditional intensity measure approach and holistic ground-motion selection. Earthquake Engineering & Structural Dynamics39(12): 1321–1342.
13.
BradleyBA (2011) Design seismic demands from seismic response analyses: A probability-based approach. Earthquake Spectra27(1): 213–224.
14.
Building Seismic Safety Council (2003) National earthquake hazards reduction program recommended provisions and commentary for seismic regulations for new buildings and other structures (FEMA 450), part 1: Provisions. Washington, DC: Federal Emergency Management Agency, pp. 356.
15.
BurattiNStaffordPJBommerJJ (2011) Earthquake accelerogram selection and scaling procedures for estimating the distribution of drift response. Journal of Structural Engineering137(3): 345–357.
16.
California Department of Transportation (2018) Seismic design criteria v2.0. Sacramento CA: California Department of Transportation.
17.
ChandramohanRBakerJWDeierleinGG (2016) Impact of hazard-consistent ground motion duration in structural collapse risk assessment. Earthquake Engineering & Structural Dynamics45(8): 1357–1379.
18.
CordovaPPDeierleinGGMehannySSCornellCA (2000) Development of a two-parameter seismic intensity measure and probabilistic assessment procedure. In: The second U.S.-Japan workshop on performance-based earthquake engineering methodology for reinforced concrete building structures, Sapporo, Japan, 11–13 September.
19.
CornellCAJalayerFHamburgerROFoutchDA (2002) Probabilistic basis for 2000 SAC federal emergency management agency steel moment frame guidelines. Journal of Structural Engineering128(4): 526–533.
20.
DahalLBurtonHOnyambuS (2022) Quantifying the effect of probability model misspecification in seismic collapse risk assessment. Structural Safety96: 102185.
21.
DastmalchiSBurtonHV (2021) Effect of modeling uncertainty on multi-limit state performance of controlled rocking steel braced frames. Journal of Building Engineering39: 102308.
22.
DhaeneGHoorelbekeD (2004) The information matrix test with bootstrap-based covariance matrix estimation. Economics Letters82(3): 341–347.
23.
EadsLMirandaEKrawinklerHLignosDG (2013) An efficient method for estimating the collapse risk of structures in seismic regions. Earthquake Engineering & Structural Dynamics42(1): 25–41.
24.
EfronB (2000) The bootstrap and modern statistics. Journal of the American Statistical Association95(452): 1293–1296.
25.
Federal Emergency Management Agency (2012) FEMA P-58-1: Seismic Performance Assessment of Buildings Volume 1—Methodology. Washington, DC: Federal Emergency Management Agency.
26.
FieldEHArrowsmithRJBiasiGPBirdPDawsonTEFelzerKRJacksonDDJohnsonKMJordanTHMaddenCMichaelAJMilnerKRPageMTParsonsTPowersPMShawBEThatcherWRWeldonRJIIZengandY (2014) Uniform California earthquake rupture forecast, version 3 (UCERF3)—The time-independent model. Bulletin of the Seismological Society of America104(3): 1122–1180.
27.
GehlPDouglasJSeyediDM (2015) Influence of the number of dynamic analyses on the accuracy of structural response estimates. Earthquake Spectra31(1): 97–113.
28.
HaindlMBurtonHSattarS (2023) Quantification of equivalent strut modeling uncertainty and its effects on the seismic performance of masonry infilled reinforced concrete frames. Journal of Earthquake Engineering28: 41–61.
29.
IbarraLF (2004) Global Collapse of Frame Structures under Seismic Excitations. PhD Thesis, Stanford University, Stanford, CA.
30.
ImbensGWKolesarM (2016) Robust standard errors in small samples: Some practical advice. Review of Economics and Statistics98(4): 701–712.
31.
JalayerFCornellC (2009) Alternative non-linear demand estimation methods for probability-based seismic assessments. Earthquake Engineering & Structural Dynamics38(8): 951–972.
32.
JalayerFEbrahimianHMianoAManfrediGSezenH (2017) Analytical fragility assessment using unscaled ground motion records. Earthquake Engineering & Structural Dynamics46(15): 2639–2663.
33.
JalayerFEbrahimianHTrevlopoulosKBradleyB (2023) Empirical tsunami fragility modelling for hierarchical damage levels. Natural Hazards and Earth System Sciences23(2): 909–931.
34.
JayaramNBakerJW (2008) Statistical tests of the joint distribution of spectral acceleration values. Bulletin of the Seismological Society of America98(5): 2231–2243.
35.
KianiJCampCPezeshkS (2018) On the number of required response history analyses. Bulletin of Earthquake Engineering16: 5195–5226.
36.
KingGRobertsME (2015) How robust standard errors expose methodological problems they do not fix, and what to do about it. Political Analysis23(2): 159–179.
37.
KowalskyMJPriestleyMNMacRaeGA (1995) Displacement-based design of RC bridge columns in seismic regions. Earthquake Engineering & Structural Dynamics24(12): 1623–1643.
38.
LallemantDKiremidjianA (2017) Accounting for uncertainty in earthquake fragility curves. In: 16th world conference on earthquake engineering 16WCEE, Santiago, Chile, 9–13 January.
LiSHedayati DezfuliFWangJAlamMS (2020) Performance-based seismic loss assessment of isolated simply-supported highway bridges retrofitted with different shape memory alloy cable restrainers in a life-cycle context. Journal of Intelligent Material Systems and Structures31(8): 1053–1075.
41.
MangalathuS (2017) Performance Based Grouping and Fragility Analysis of Box-Girder Bridges in California. PhD Thesis, Georgia Institute of Technology, Atlanta, GA.
42.
MutoMBeckJL (2008) Bayesian updating and model class selection for hysteretic structural models using stochastic simulation. Journal of Vibration and Control14(1): 27–34.
43.
ÖzsaraçV (2023) EzGM: Toolbox for ground motion record selection and processing. Available at: https://github.com/volkanozsarac/EzGM (accessed 3 June 2023).
44.
PaganiMMonelliDWeatherillGDanciuLCrowleyHSilvaVHenshawPButlerLNastasiMPanzeriLSimionatoMViganoD (2014) OpenQuake engine: An open hazard (and risk) software for the global earthquake model. Seismological Research Letters85(3): 692–702.
45.
PetersenMDMoschettiMPPowersPMMuellerCSHallerKMFrankelAZengYRezaeianSHarmsenSBoydOSFieldEHChenRRukstalesKSLucoNWheelerRWilliamsROlsenAH (2015) The 2014 United States national seismic hazard model. Earthquake Spectra31(1. Suppl.): S1–S30.
46.
SáezELopez-CaballeroFModaressi-Farahmand-RazaviA (2011) Effect of the inelastic dynamic soil–structure interaction on the seismic vulnerability assessment. Structural Safety33(1): 51–63.
47.
StaffordPJBooreDMYoungsRRBommerJJ (2022) Host-region parameters for an adjustable model for crustal earthquakes to facilitate the implementation of the backbone approach to building ground-motion logic trees in probabilistic seismic hazard analysis. Earthquake Spectra38(2): 917–949.
48.
VamvatsikosDCornellCA (2005) Developing efficient scalar and vector intensity measures for IDA capacity estimation by incorporating elastic spectral shape information. Earthquake Engineering & Structural Dynamics34(13): 1573–1600.
49.
WhiteH (1982) Maximum likelihood estimation of misspecified models. Econometrica: Journal of the Econometric Society50(1): 1–25.
50.
WilcoxRR (1981) A review of the beta-binomial model and its extensions. Journal of Educational Statistics6(1): 3–32.
51.
ZahnFParkRPriestleyMChapmanH (1986) Development of design procedures for the flexural strength and ductility of reinforced concrete bridge columns. Bulletin of the New Zealand Society for Earthquake Engineering19(3): 200–212.
52.
ZhuMMcKennaFScottMH (2018) OpenSeesPy: Python library for the OpenSees finite element framework. Softwarex7: 6–11.
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.