-
**Cohort definition and sample size reporting are internally inconsistent across the manuscript, making it unclear which animals contributed to which analyses and undermining reproducibility. Methods report a final cohort of $N=28$ with specific sex/colony counts (Sec. 2.1), while Abstract/Results repeatedly use $N=30$ and provide different sex/colony compositions (Sec. 3.1–3.4; Table 2; figures). $\rm DNAmAge$ descriptive statistics also differ between Methods and Results despite the same stated range (Sec. 2.1 vs Sec. 3.1).** *Recommendation:* Provide a single, unambiguous accounting of the dataset: a CONSORT-like flow from initial $N$ (e.g., $N=41$) through exclusions (with reasons) to the final $N$ for each modality and each analysis (behavior, $\rm DNAmAge$, DTI, regression). Add a table listing per-animal data availability (or at least counts of complete cases by modality). Then harmonize $N$, sex/colony counts, $\rm DNAmAge$ mean$\pm$SD, figure captions, and Table 2 degrees of freedom to match the actually analyzed cohorts (Sec. 2.1; Sec. 3.1–3.4).
-
**DTI predictors (Global_FA, Global_MD) appear physiologically implausible (e.g., global $\rm FA \approx 0.99$ with extremely low diffusivities), strongly suggesting acquisition/preprocessing/unit-scaling/masking problems. Because these variables are core predictors in the CAE regression and central to the structural-decoupling narrative, their questionable validity jeopardizes the main conclusions (Sec. 2.3; Sec. 3.3; Sec. 3.4.1; Sec. 3.5).** *Recommendation:* Add (and act on) a focused DTI QA/QC section. Concretely: (i) explicitly state units and scaling for MD/AD/RD ($\rm mm^2/s$ vs $m^2/s$; whether $1e^{-3}$ scaling is applied), (ii) describe brain/WM mask generation and exactly how “global mean” metrics were computed (whole-brain vs WM-only vs atlas ROIs vs skeletonized voxels), (iii) report preprocessing steps and settings (eddy/motion/susceptibility correction; outlier handling) and software/versions (Sec. 2.3), and (iv) provide basic QC visualizations (representative FA/MD maps $+$ masks; FA/MD histograms across voxels; motion/SNR summaries). If an error is found, recompute global metrics and update all downstream analyses. If uncertainty remains, move structure-based conclusions to exploratory/sensitivity analyses and foreground results that do not depend on DTI integrity (e.g., $\rm CAE\sim DNAmAge$; Sec. 3.4.1).
-
**CAE is central to the paper but is not specified with a fully explicit, unambiguous mathematical definition, and current edge-case handling can bias CAE upward. The displayed formula is ambiguous without parentheses and may contain a sign/precedence interpretation that would incorrectly increase CAE with higher LTM perseveration. Additionally, perseverative error rate is set to 0 when total entries in a phase are 0 (division-by-zero handling), which treats non-participation as perfect performance and can inflate CAE; small denominators also make rates unstable (Sec. 2.2.1–2.2.2; Sec. 3.2).** *Recommendation:* In Sec. 2.2.2, provide the exact CAE formula with explicit parentheses, signs, weighting (STM vs LTM), and theoretical range; include a worked example. Verify the typeset equation matches the implemented code (and fix if not). Replace the division-by-zero convention: treat phases with 0 entries as missing/undefined (or enforce a minimum-entry threshold), and report how many bats/phases are affected. Add sensitivity analyses for CAE (e.g., excluding low-engagement phases; alternative thresholds; treating STM/LTM separately; or modeling perseverative counts using an appropriate binomial framework with exposure/denominator) and report whether key conclusions persist (Sec. 3.2; Sec. 3.4).
-
**The manuscript over-interprets a non-significant, low-$R^2$ regression as “direct evidence” of decoupling. With $N\approx 30$, multiple predictors, potential measurement noise (behavior and especially DTI), bounded/skewed outcome, and possible nonlinearities, failure to reject the null is not strong evidence of absence of association (Sec. 3.4.1; Sec. 3.5; Abstract; Conclusion).** *Recommendation:* Reframe claims in the Abstract, Sec. 3.5, and Conclusion to emphasize that the study did not detect strong linear associations under the current design and power, rather than asserting direct evidence of decoupling. Report effect sizes with confidence intervals (and standardized coefficients) for all predictors, including the borderline $\rm DNAmAge$ effect noted in Sec. 3.4.1. Add a power/sensitivity analysis (minimum detectable effect size given $N$) and/or Bayesian estimation to quantify evidence for near-zero effects. Include simpler supporting analyses (univariate $\rm CAE\sim DNAmAge$; $\rm CAE\sim FA$; $\rm CAE\sim MD$; partial correlations) and clearly distinguish “absence of evidence” from “evidence of absence.”
-
**CSDI is presented as an individualized “resilience/decoupling” index, but its validity and robustness are not established and it likely inherits instability from the weak/possibly misspecified model and questionable DTI predictors. Residuals can largely reflect noise, collinearity, outliers, or model choice rather than a stable trait-like resilience measure (Sec. 2.4.2; Sec. 3.4.2).** *Recommendation:* Treat CSDI explicitly as exploratory unless you can demonstrate robustness. Add stability checks: bootstrap the regression and report uncertainty for each CSDI (or at least rank stability), and compare CSDI values across alternative specifications (e.g., $\rm CAE\sim DNAmAge+Sex$ only; with/without FA/MD; robust regression). Report whether high/low CSDI individuals remain consistent across models. Also test associations between CSDI and potential nuisance variables (total entries/engagement, colony, session/batch) to ensure CSDI is not primarily capturing these factors (Sec. 3.4.2; Sec. 3.1–3.2).
-
**Potential confounding/hierarchical structure is insufficiently addressed. Colony (Aseret vs Herzeliya) is reported but not modeled; it may proxy environmental differences, handling, scanning sessions/batch effects, or task exposure differences. Engagement (total entries) may also confound CAE if activity levels vary substantially across individuals (Sec. 2.1; Sec. 3.1–3.2; Sec. 2.2.2).** *Recommendation:* At minimum, include colony as a covariate in the CAE model (or justify exclusion given power). Report descriptive comparisons by colony/sex for $\rm DNAmAge$, CAE, and (if valid) DTI metrics (Sec. 3.1–3.2). Evaluate whether CAE correlates with total entries (and consider including total entries as a covariate or switching to a count-based model of perseveration with appropriate exposure). If methylation age or imaging were processed in batches, report and (if possible) adjust for batch/scanning session effects (Sec. 2.1; Sec. 2.3; Sec. 2.4.2).
-
**Aims and narrative emphasize regional “Structural Preservation Hotspots” and region-wise neural signatures, but the Results state regional analyses could not be performed due to technical issues. This creates a mismatch between promised contributions and delivered results (Introduction/Sec. 1; Sec. 2.4.1–2.4.3; Sec. 3.3).** *Recommendation:* Refocus the manuscript around what was actually executed (global metrics $+$ behavioral/$\rm DNAmAge$ analyses) and clearly label regional hotspot mapping as planned/future work. Move unexecuted regional methods (Sec. 2.4.1; Sec. 2.4.3) to an Appendix or a dedicated “Planned analyses” subsection, and update the Abstract/Introduction/Conclusion so they do not imply regional neural signatures were identified in this study (Sec. 3.3; Sec. 3.5).
-
**Modeling and reporting choices are not well matched to the data characteristics and currently hinder interpretability: CAE is bounded in $[0,1]$ and appears skewed/near-ceiling, yet OLS is used without adequate diagnostics; the regression output/table appears corrupted and/or inconsistent with described software; and numerical stability/collinearity concerns are suggested (e.g., extremely large condition number mentioned in the unstructured report) but not addressed (Sec. 3.4.1; Figure 4; Figure 6; Sec. 2.4.2).** *Recommendation:* Improve statistical reporting and align the model to the outcome. Provide a clean regression table (coefficients, SEs, $t$, exact $p$, CIs, $R^2$/$\rm adj$-$R^2$, $F$ and $p$, AIC/BIC) and clarify the software used (R vs Python/statsmodels) with package versions (Sec. 3.4.1; Sec. 2.4). Add diagnostics: residual plots, influence, and multicollinearity checks (VIF/condition number), and standardize continuous predictors for interpretability. Consider a model appropriate for bounded outcomes (beta regression; quasi-binomial on error rates; or robust/Spearman-based analyses) and report whether conclusions are consistent across approaches (Figure 4; Sec. 3.4).