On the comparability between studies in predictive ecotoxicology
OPEN ACCESS
Loading...
Author / Producer
Date
2025-09
Publication Type
Review Article
ETH Bibliography
yes
OPEN ACCESS
Data
Rights / License
Abstract
Comparability across in silico predictive ecotoxicology studies remains a significant challenge, particularly when assessing model performance. In this work, we identify key criteria necessary for meaningful comparison between independent studies: (i) the use of identical datasets that represent the same chemical and/or taxonomic space; (ii) consistent data cleaning procedures; (iii) identical train/test splits; (iv) clearly defined evaluation metrics, as subtle differences — such as alternative formulations of R² — can lead to misleading discrepancies; and (v) transparent reporting through code and dataset sharing. Our review of recent literature on fish acute toxicity prediction reveals a critical gap: no two studies fully meet these criteria, rendering cross-study comparisons unreliable. This lack of comparability hampers scientific progress in the field. To address this, we advocate for the adoption of benchmark datasets with standardized cleaning protocols, version control, and defined data splits. We further emphasize the importance of precise metric definitions and transparent reporting practices, including code availability and the use of structured reporting or data sheets, to foster reproducibility and advance the discipline.
Permanent link
Publication status
published
External links
Editor
Book title
Journal / series
Volume
35
Pages / Article No.
100367
Publisher
Elsevier
Event
Edition / version
Methods
Software
Geographic location
Date collected
Date created
Subject
Machine learning; Quantitative Structure Activity; Relationship-QSAR; Quantitative Structure Toxicity; Relationship-QSTR; Reproducibility; Performance metrics; Benchmark; Environmental hazard assessment
Organisational unit
Notes
Funding
101057014/22.00417 - Partnership for the Assessment of Risks from Chemicals (SBFI)
