Fit Measures in SmartPLS
SmartPLS offers the following fit measures:
 SRMR
 Exact fit criteria d_ULS and d_G
 NFI
 Chi²
For the approximate fit indices such as SRMR and NFI, you may directly look at the outcomes of a PLSSEM or PLScSEM model estimation (i.e., the results report) and these criteria's values with a certain threshold (e.g., SRMR < 0.08 and NFI > 0.90).
For the bootstrapbased test for the exact overall model fit measures d_ULS and d_G you may consider the inference statistics for an assessment. Therefore, you need to run the bootstrap procedure and to use the “complete bootstrap” option in SmartPLS. When running the bootstrap procedure, you will notice that the procedure counts two times up to the specified number of bootstrapping samples:
 In the first round, SmartPLS uses the standard bootstrapping procedure to get the inference statistics for the model parameters (e.g., path coefficients, weights, etc.).
 In the second round, SmartPLS uses an adapted BollenStine bootstrapping procedure as described in Dijkstra and Henseler (2015; also see Bollen and Stine, 1992; Yuan and Hayashi, 2003) to create confidence intervals for the d_ULS, d_G, and SRMR criteria (note that SmartPLS has two computation runs in the second round: one for the saturated model and one for the estimated model).
Note: Also see the information on goodness of fit (GoF), which researchers should not use.
Standardized Root Mean Square Residual (SRMR)
While the root mean square residual (RMSR) is a measure of the mean absolute value of the covariance residuals, the standardized root mean square residual (SRMR) based on transforming both the sample covariance matrix and the predicted covariance matrix into correlation matrices. Note: Literature on PLSSEM needs to better explain where and how the covariance matrix is derived in PLSSEM (since it is different from CBSEM, which is a full information method and PLSSEM is not). Most important, should the researcher use the estimated model (most reasonable choice) or the saturated model to obtain the covariance matrix.
The SRMR is defined as the difference between the observed correlation and the model implied correlation matrix. Thus, it allows assessing the average magnitude of the discrepancies between observed and expected correlations as an absolute measure of (model) fit criterion.
A value less than 0.10 or of 0.08 (in a more conservative version; see Hu and Bentler, 1999) are considered a good fit. Henseler et al. (2014) introduce the SRMR as a goodness of fit measure for PLSSEM that can be used to avoid model misspecification.
SmartPLS also provides bootstrapbased inference statistics of the SRMR criterion. For the interpretation of SRMR bootstrap confidence interval results see the bootstrapbased test for the exact overall model fit.
Bootstrapbased test for the exact overall model fit
The bootstrapbased test for the exact overall model fit tests the statistical (bootstrapbased) inference of the discrepancy between the empirical covariance matrix and the covariance matrix implied by the composite factor model. As defined by Dijkstra and Henseler (2015; also see Schuberth et al., 2022), d_ULS (i.e., the squared Euclidean distance) and d_G (i.e., the geodesic distance) represent two different ways to compute this discrepancy. The BollenStine (1992) bootstrap routine provides the confidence intervals of these discrepancy values. The d_G criterion builds on PLSSEM eigenvalue computations. However, the question remains how these eigenvalues differ from CBSEM.
Note: The value of the d_ULS and d_G in itself do not pertain any value. Only the BollenStine bootstrapping results of the bootstrapbased test for the exact overall model fit, as provided my SmartPLS, allow an interpretation of results. More specifically, since the d_ULS and d_G (and SRMR) confidence intervals are not obtained by running the regular bootstrapping procedure, but the adapted BollenStine bootstrapping procedure, their results interpretation somewhat differs from the regular bootstrap outcomes.
For the bootstrapbased test for the exact overall model fit (i.e., d_ULS and d_G), you compare their original value against the confidence interval created from the sampling distribution. The confidence interval should include the original value. Hence, the upper bound of the confidence interval should be larger than the original value of the d_ULS and d_G fit criteria to indicate that the model has a “good fit”. Choose the confidence interval in a way that the upper bound is at the 95% or 99% point.
In other words, a model fits well if the difference between the correlation matrix implied by your model and the empirical correlation matrix is so small that it can be purely attributed to sampling error. Hence, the difference between the correlation matrix implied by your model and the empirical correlation matrix should be nonsignificant (p > 0.05). Otherwise, if the discrepancy is significant (p < 0.05), model fit has not been established.
Normed Fit Index (NFI) or Bentler and Bonett Index
One of the first fit measures proposed in the SEM literature is the normed fit index by Bentler and Bonett (1980). It computes the Chi² value of the proposed model and compares it against a meaningful benchmark. Since the Chi² value of the proposed model in itself does not provide sufficient information to judge model fit, the NFI uses the Chi² value from the null model, as a yardstick. Literature, however, does not explain how the PLSSEM Chi² value differs from the CBSEM one.
The NFI is then defined as 1 minus the Chi² value of the proposed model divided by the Chi² values of the null model. Consequently, the NFI results in values between 0 and 1. The closer the NFI to 1, the better the fit. NFI values above 0.9 usually represent acceptable fit. Lohmöller (1989) provides detailed information on the NFI computation of PLS path models. However, for the applied user, these explications are quite difficult to comprehend.
The NFI represents an incremental fit measure. As such, a major disadvantage is that it does not penalize for model complexity. The more parameters in the model, the larger (i.e., better) the NFI result. It is for this reason that this measure is not recommended, but alternatives such as the nonnormed fit index (NNFI) or TuckerLewis index, which penalizes the Chi² values by the degrees of freedom (df). Lohmöller (1989) suggests computing the NNFI of PLS path models. However, the NNFI has not been implemented in SmartPLS, yet.
Chi² and Degrees of Freedom
Assuming a multinormal distribution, the Chi² value of a PLS path model with df degrees of freedom approximately is (N1)*L, whereby N is the number of observations and L the maximum likelihood function as defined by Lohmöller (1989). The degrees of freedom (df) are defined as (K² + K) /2 – t, whereby is the number of manifest variables in the PLS path model and t the number of independent variables to estimate the model implied covariance matrix. However, future research must clearly define how to determine the degrees of freedom of composite models, common factor models, and mixed models when using PLSSEM.
Estimated and Saturated Model
The distinction of estimated and saturated models in PLSSEM is in its very early stages. Future research must provide detailed explanations and recommendations on the computation, usage and interpretation of these outcomes.
The saturated model assesses correlation between all constructs. The estimated model is a model which is based on a total effect scheme and takes the model structure into account. It is hence a more restricted version of the fit measure.
Researchers often struggle to choose between the estimated and saturated model when trying to report the fir of a PLS path model. At this stage, PLSSEM literature is very vague on the use of fit criteria in general and, in specific, the choice between the estimated and saturated model. However, the estimated model seems to be a reasonable choice, if a researcher makes the questionable decision to report the fit results of the PLS path model.
Composite Model Fit Measures
If you like to obtain the composite model fit measures, use formative measurement models for all constructs in the PLS path model. After model estimation, refer to the estimated (or saturated?) model outcomes.
Common Factor Model Fit Measures
If you like to obtain the common factor model fit measures, use reflective measurement models for all constructs in the PLS path model. After model estimation, refer to the estimated (or saturated?) model SRMR outcomes. However, when assuming common factor models for all constructs in the PLS path model the question remains, why the researcher does not use CBSEM in the first place to estimate and evaluate such a model.
Mixed Model Fit Measures
If you use both reflective and formative measurement models, SmartPLS 3.2.4 (and subsequent versions) provides the mixed model fit measures considering common factor models for reflective measurement models and composite models for formative measurement models. However, at this stage, PLSSEM literature does not provide much support, why and how a researcher would theoretically distinguish between constructs represented by common factors and composites in the same model, the need for their mixed use, and the necessity to report the model fit.
Note on Model Fit in PLSSEM
Lohmöller (1989) already offers a set of fit measures. But he states that they have been introduced to provide a comparison to LISREL results rather than to represent an appropriate PLSSEM index. More specifically, Lohmöller (1989) finds that some fit measures imply restrictive assumptions on the residual covariances, which PLSSEM does not imply when estimating the model. For example, certain fit measures assume a common factor model, which requires uncorrelated outer residuals. In contrast, the outer residuals of composite models are not required to be uncorrelated. Hence, they are inappropriate for PLSSEM.
However, when mimicking CBSEM models with the consistent PLS (PLScSEM) approach, one also mimics common factor models with the PLSSEM approach. Hence, when using PLScSEM for a path model that only includes reflectively measured constructs (i.e., common factor models), one may be interested in the model fit. Thereby, it is more comprehensively possible to mimic CBSEM via the PLScSEM approach or to compare the results from the two approaches. Against this background Sarstedt et al. (2017, 2021) conclude "validation using goodnessoffit measures is also relevant in a PLSSEM context but less so compared to factorbased SEM".
References

Bollen, K. A., & Stine, R. A. (1992). Bootstrapping GoodnessofFit Measures in Structural Equation Models. Sociological Methods & Research, 21(2): 205229.

Bentler, P. M., & Bonett, D. G. (1980). Significance Tests and GoodnessofFit in the Analysis of Covariance Structures, Psychological Bulletin, 88: 588600.

Dijkstra, T. K. and Henseler, J. (2015). Consistent and Asymptotically Normal PLS Estimators for Linear Structural Equations, Computational Statistics & Data Analysis, 81(1): 1023.

Hair, J. F., Hollingsworth, C. L., Randolph, A. B., and Chong, A. Y. L (2017). An Updated and Expanded Assessment of PLSSEM in Information Systems Research. Industrial Management & Data Systems, 117(3): 442458.

Hair, J. F., Hult, G. T. M., Ringle, C. M., & Sarstedt, M. (2022). A Primer on Partial Least Squares Structural Equation Modeling (PLSSEM) (3 ed.). Thousand Oaks, CA: Sage.

Henseler, J., Dijkstra, T. K., Sarstedt, M., Ringle, C. M., Diamantopoulos, A., Straub, D. W., Ketchen, D. J., Hair, J. F., Hult, G. T. M., and Calantone, R. J. 2014. Common Beliefs and Reality about Partial Least Squares: Comments on Rönkkö & Evermann (2013), Organizational Research Methods, 17(2): 182209.

Hu, L.t., and Bentler, P. M. (1998). Fit Indices in Covariance Structure Modeling: Sensitivity to Underparameterized Model Misspecification, Psychological Methods, 3(4): 424453.

Lohmöller, J.B. (1989). Latent Variable Path Modeling with Partial Least Squares, Physica: Heidelberg.

Sarstedt, M., Ringle, C. M., & Hair, J. F. (2017). Partial Least Squares Structural Equation Modeling. In C. Homburg, M. Klarmann, & A. E. Vomberg (Eds.), Handbook of Market Research (pp. 1–40). Springer: Cham.

Sarstedt, M., Ringle, C. M., & Hair, J. F. (2021). Partial Least Squares Structural Equation Modeling. In C. Homburg, M. Klarmann, & A. E. Vomberg (Eds.), Handbook of Market Research (pp. 1–47). Springer: Cham.

Schuberth, F., Rademaker, M. E., & Henseler, J. (2023). Assessing the Overall Fit of Composite Models Estimated by Partial Least Squares Path Modeling. European Journal of Marketing, 57(6), 16781702.
Please always cite the use of SmartPLS!
Ringle, Christian M., Wende, Sven, & Becker, JanMichael. (2024). SmartPLS 4. Bönningstedt: SmartPLS. Retrieved from https://www.smartpls.com