CRAN Package Check Results for Package parameters

Last updated on 2026-02-11 11:51:01 CET.

Flavor Version Tinstall Tcheck Ttotal Status Flags
r-devel-linux-x86_64-debian-clang 0.28.3 OK
r-devel-linux-x86_64-debian-gcc 0.28.3 20.52 259.80 280.32 OK
r-devel-linux-x86_64-fedora-clang 0.28.3 57.00 684.38 741.38 OK
r-devel-linux-x86_64-fedora-gcc 0.28.3 1.00 5.36 6.36 ERROR
r-devel-macos-arm64 0.28.3 7.00 76.00 83.00 ERROR
r-devel-windows-x86_64 0.28.3 33.00 316.00 349.00 ERROR
r-patched-linux-x86_64 0.28.3 34.83 381.23 416.06 OK
r-release-linux-x86_64 0.28.3 27.90 381.66 409.56 OK
r-release-macos-arm64 0.28.3 8.00 105.00 113.00 OK
r-release-macos-x86_64 0.28.3 24.00 515.00 539.00 OK
r-release-windows-x86_64 0.28.3 32.00 329.00 361.00 OK
r-oldrel-macos-arm64 0.28.3 9.00 101.00 110.00 NOTE
r-oldrel-macos-x86_64 0.28.3 26.00 801.00 827.00 OK
r-oldrel-windows-x86_64 0.28.3 47.00 435.00 482.00 NOTE

Check Details

Version: 0.28.3
Check: whether package can be installed
Result: ERROR Installation failed. Flavor: r-devel-linux-x86_64-fedora-gcc

Version: 0.28.3
Check: tests
Result: ERROR Running ‘testthat.R’ [37s/21s] Running the tests in ‘tests/testthat.R’ failed. Complete output: > library(parameters) > library(testthat) > > test_check("parameters") Starting 2 test processes. > test-format_model_parameters.R: Cannot compute confidence intervals for random effects parameters. > test-format_model_parameters.R: Your model may suffer from singularity (see `?lme4::isSingular` and > test-format_model_parameters.R: `?performance::check_singularity`). Saving _problems/test-format_model_parameters-190.R > test-include_reference.R: Your model may suffer from singularity (see `?lme4::isSingular` and > test-include_reference.R: `?performance::check_singularity`). > test-include_reference.R: Some of the confidence intervals of the random effects parameters are > test-include_reference.R: probably not meaningful! > test-include_reference.R: You may try to impose a prior on the random effects parameters, e.g. > test-include_reference.R: using the glmmTMB package. > test-include_reference.R: Your model may suffer from singularity (see `?lme4::isSingular` and > test-include_reference.R: `?performance::check_singularity`). > test-include_reference.R: Some of the confidence intervals of the random effects parameters are > test-include_reference.R: probably not meaningful! > test-include_reference.R: You may try to impose a prior on the random effects parameters, e.g. > test-include_reference.R: using the glmmTMB package. > test-model_parameters.afex_aov.R: Contrasts set to contr.sum for the following variables: condition, talk > test-model_parameters.afex_aov.R: Contrasts set to contr.sum for the following variables: condition, talk > test-model_parameters.afex_aov.R: Contrasts set to contr.sum for the following variables: treatment, gender > test-n_factors.R: [1] "# Method Agreement Procedure:" > test-n_factors.R: [2] "" > test-n_factors.R: [3] "The choice of 1 dimensions is supported by 11 (84.62%) methods out of 13 (Bartlett, Anderson, Lawley, Optimal coordinates, Acceleration factor, Parallel analysis, Kaiser criterion, Scree (SE), Scree (R2), VSS complexity 1, Velicer's MAP)." > test-n_factors.R: [1] "# Method Agreement Procedure:" > test-n_factors.R: [2] "" > test-n_factors.R: [3] "The choice of 1 dimensions is supported by 3 (60.00%) methods out of 5 (Velicer's MAP, BIC, BIC (adjusted))." [ FAIL 1 | WARN 0 | SKIP 131 | PASS 713 ] ══ Skipped tests (131) ═════════════════════════════════════════════════════════ • On CRAN (118): 'test-Hmisc.R:1:1', 'test-GLMMadaptive.R:1:1', 'test-averaging.R:1:1', 'test-backticks.R:1:1', 'test-bootstrap_emmeans.R:1:1', 'test-bootstrap_parameters.R:1:1', 'test-brms.R:1:1', 'test-compare_parameters.R:90:5', 'test-compare_parameters.R:95:5', 'test-complete_separation.R:4:1', 'test-complete_separation.R:18:1', 'test-complete_separation.R:28:1', 'test-coxph.R:69:1', 'test-efa.R:1:1', 'test-emmGrid-df_colname.R:1:1', 'test-equivalence_test.R:3:1', 'test-equivalence_test.R:13:1', 'test-equivalence_test.R:22:3', 'test-equivalence_test.R:112:3', 'test-factor_analysis.R:2:3', 'test-factor_analysis.R:124:3', 'test-format_model_parameters2.R:2:3', 'test-gam.R:30:1', 'test-get_scores.R:1:1', 'test-glmer.R:1:1', 'test-glmmTMB-2.R:1:1', 'test-glmmTMB-profile_CI.R:2:3', 'test-glmmTMB.R:1:1', 'test-helper.R:1:1', 'test-ivreg.R:45:1', 'test-lcmm.R:1:1', 'test-lmerTest.R:1:1', 'test-include_reference.R:4:1', 'test-include_reference.R:62:1', 'test-include_reference.R:110:1', 'test-mipo.R:5:1', 'test-mipo.R:23:1', 'test-mmrm.R:1:1', 'test-model_parameters.BFBayesFactor.R:4:3', 'test-model_parameters.BFBayesFactor.R:77:3', 'test-model_parameters.BFBayesFactor.R:114:3', 'test-model_parameters.anova.R:1:1', 'test-model_parameters.aov.R:1:1', 'test-model_parameters.aov_es_ci.R:183:3', 'test-model_parameters.aov_es_ci.R:294:3', 'test-model_parameters.aov_es_ci.R:344:3', 'test-model_parameters.aov_es_ci.R:397:3', 'test-model_parameters.bracl.R:1:1', 'test-model_parameters.cgam.R:1:1', 'test-model_parameters.coxme.R:1:1', 'test-marginaleffects.R:170:1', 'test-marginaleffects.R:199:3', 'test-model_parameters.epi2x2.R:1:1', 'test-model_parameters.efa_cfa.R:30:3', 'test-model_parameters.fixest_multi.R:1:1', 'test-model_parameters.fixest.R:2:3', 'test-model_parameters.fixest.R:77:3', 'test-model_parameters.fixest.R:145:1', 'test-model_parameters.glmgee.R:1:1', 'test-model_parameters.glm.R:35:1', 'test-model_parameters.glm.R:67:1', 'test-model_parameters.logistf.R:1:1', 'test-model_parameters.logitr.R:1:1', 'test-model_parameters.mclogit.R:1:1', 'test-model_parameters.mediate.R:1:1', 'test-model_parameters.mixed.R:2:1', 'test-model_parameters.nnet.R:5:1', 'test-model_parameters_df.R:1:1', 'test-model_parameters.vgam.R:3:1', 'test-model_parameters_ordinal.R:1:1', 'test-model_parameters_random_pars.R:1:1', 'test-model_parameters_std.R:1:1', 'test-model_parameters_std_mixed.R:1:1', 'test-n_factors.R:10:3', 'test-n_factors.R:26:3', 'test-n_factors.R:76:3', 'test-p_adjust.R:1:1', 'test-p_direction.R:1:1', 'test-p_significance.R:1:1', 'test-p_value.R:14:1', 'test-panelr.R:1:1', 'test-pipe.R:1:1', 'test-pca.R:64:1', 'test-polr.R:1:1', 'test-plm.R:97:1', 'test-posterior.R:1:1', 'test-pool_parameters.R:1:1', 'test-pool_parameters.R:32:1', 'test-pretty_names.R:40:1', 'test-pretty_names.R:73:5', 'test-print_AER_labels.R:5:1', 'test-printing-stan.R:1:1', 'test-printing.R:1:1', 'test-quantreg.R:1:1', 'test-random_effects_ci-glmmTMB.R:3:1', 'test-random_effects_ci.R:1:1', 'test-robust.R:1:1', 'test-serp.R:5:1', 'test-printing2.R:14:5', 'test-printing2.R:21:5', 'test-printing2.R:26:5', 'test-printing2.R:31:5', 'test-printing2.R:36:5', 'test-printing2.R:48:5', 'test-printing2.R:91:7', 'test-printing2.R:126:5', 'test-svylme.R:1:1', 'test-svyolr.R:1:1', 'test-visualisation_recipe.R:1:1', 'test-wrs2.R:55:1', 'test-standardize_parameters.R:28:1', 'test-standardize_parameters.R:36:3', 'test-standardize_parameters.R:61:3', 'test-standardize_parameters.R:173:3', 'test-standardize_parameters.R:297:3', 'test-standardize_parameters.R:332:3', 'test-standardize_parameters.R:425:3', 'test-standardize_parameters.R:515:3' • On Mac (7): 'test-group_level_total.R:1:1', 'test-nestedLogit.R:78:3', 'test-rstanarm.R:1:1', 'test-sampleSelection.R:1:1', 'test-simulate_model.R:1:1', 'test-simulate_parameters.R:1:1', 'test-weightit.R:1:1' • TODO: check this test locally, fails on CI, probably due to scoping issues? (1): 'test-marginaleffects.R:280:3' • TODO: fix this test (1): 'test-model_parameters.lqmm.R:40:3' • TODO: this one actually is not correct. (1): 'test-model_parameters_robust.R:127:3' • empty test (2): 'test-wrs2.R:69:1', 'test-wrs2.R:81:1' • {performance} is not installed (1): 'test-dominance_analysis.R:1:1' ══ Failed tests ════════════════════════════════════════════════════════════════ ── Error ('test-format_model_parameters.R:190:3'): format, compare_parameters, mixed models ── Error in ``[.data.frame`(out, c("Parameter", "Level", "Coefficient", "SE", ci_cols, stat_column, "df_error", "p", "Effects", "Group"))`: undefined columns selected Backtrace: ▆ 1. └─parameters::compare_parameters(...) at test-format_model_parameters.R:190:3 2. └─base::lapply(...) 3. └─parameters (local) FUN(X[[i]], ...) 4. ├─parameters::model_parameters(...) 5. └─parameters:::model_parameters.glmmTMB(...) 6. └─parameters:::.add_random_effects_glmmTMB(...) 7. ├─parameters:::.extract_random_variances(...) 8. └─parameters:::.extract_random_variances.glmmTMB(...) 9. ├─base::suppressWarnings(...) 10. │ └─base::withCallingHandlers(...) 11. └─parameters:::.extract_random_variances_helper(...) 12. ├─...[] 13. └─base::`[.data.frame`(...) [ FAIL 1 | WARN 0 | SKIP 131 | PASS 713 ] Error: ! Test failures. Execution halted Flavor: r-devel-macos-arm64

Version: 0.28.3
Check: tests
Result: ERROR Running 'testthat.R' [54s] Running the tests in 'tests/testthat.R' failed. Complete output: > library(parameters) > library(testthat) > > test_check("parameters") Starting 2 test processes. > test-include_reference.R: Your model may suffer from singularity (see `?lme4::isSingular` and > test-include_reference.R: `?performance::check_singularity`). > test-include_reference.R: Some of the confidence intervals of the random effects parameters are > test-include_reference.R: probably not meaningful! > test-include_reference.R: You may try to impose a prior on the random effects parameters, e.g. > test-include_reference.R: using the glmmTMB package. > test-include_reference.R: Your model may suffer from singularity (see `?lme4::isSingular` and > test-include_reference.R: `?performance::check_singularity`). > test-include_reference.R: Some of the confidence intervals of the random effects parameters are > test-include_reference.R: probably not meaningful! > test-include_reference.R: You may try to impose a prior on the random effects parameters, e.g. > test-include_reference.R: using the glmmTMB package. > test-model_parameters.afex_aov.R: Contrasts set to contr.sum for the following variables: condition, talk > test-model_parameters.afex_aov.R: Contrasts set to contr.sum for the following variables: condition, talk > test-model_parameters.afex_aov.R: Contrasts set to contr.sum for the following variables: treatment, gender > test-n_factors.R: [1] "# Method Agreement Procedure:" > test-n_factors.R: [2] "" > test-n_factors.R: [3] "The choice of 1 dimensions is supported by 11 (84.62%) methods out of 13 (Bartlett, Anderson, Lawley, Optimal coordinates, Acceleration factor, Parallel analysis, Kaiser criterion, Scree (SE), Scree (R2), VSS complexity 1, Velicer's MAP)." > test-n_factors.R: [1] "# Method Agreement Procedure:" > test-n_factors.R: [2] "" > test-n_factors.R: [3] "The choice of 1 dimensions is supported by 3 (60.00%) methods out of 5 (Velicer's MAP, BIC, BIC (adjusted))." > test-nestedLogit.R: Error: ! testthat subprocess exited in file 'test-nestedLogit.R'. Caused by error: ! R session crashed with exit code -1073741819 Backtrace: ▆ 1. └─testthat::test_check("parameters") 2. └─testthat::test_dir(...) 3. └─testthat:::test_files(...) 4. └─testthat:::test_files_parallel(...) 5. ├─withr::with_dir(...) 6. │ └─base::force(code) 7. ├─testthat::with_reporter(...) 8. │ └─base::tryCatch(...) 9. │ └─base (local) tryCatchList(expr, classes, parentenv, handlers) 10. │ └─base (local) tryCatchOne(expr, names, parentenv, handlers[[1L]]) 11. │ └─base (local) doTryCatch(return(expr), name, parentenv, handler) 12. └─testthat:::parallel_event_loop_chunky(queue, reporters, ".") 13. └─queue$poll(Inf) 14. └─base::lapply(...) 15. └─testthat (local) FUN(X[[i]], ...) 16. └─private$handle_error(msg, i) 17. └─cli::cli_abort(...) 18. └─rlang::abort(...) Execution halted Flavor: r-devel-windows-x86_64

Version: 0.28.3
Check: package dependencies
Result: NOTE Package suggested but not available for checking: ‘M3C’ Flavor: r-oldrel-macos-arm64

Version: 0.28.3
Check: package dependencies
Result: NOTE Package suggested but not available for checking: 'EGAnet' Flavor: r-oldrel-windows-x86_64

mirror server hosted at Truenetwork, Russian Federation.