Mostrar el registro sencillo del ítem
Early-fusion hybrid CNN-transformer models for multiclass ovarian tumor ultrasound classification
| dc.contributor.author | Garcia Atutxa, Igor | |
| dc.contributor.author | Martínez Más, José | |
| dc.contributor.author | Bueno Crespo, Andrés | |
| dc.contributor.author | Villanueva Flores, Francisca | |
| dc.date.accessioned | 2025-12-09T09:00:01Z | |
| dc.date.available | 2025-12-09T09:00:01Z | |
| dc.date.issued | 2025-10-15 | |
| dc.identifier.citation | Garcia-Atutxa, I., Martínez-Más, J., Bueno-Crespo, A., & Villanueva Flores, F. (2025). Early‑fusion hybrid CNN‑Transformer models for multiclass ovarian tumor ultrasound classification. Frontiers in Artificial Intelligence, 8, 1679310. | es |
| dc.identifier.uri | http://hdl.handle.net/10952/10524 | |
| dc.description.abstract | Ovarian cancer remains the deadliest gynecologic malignancy, and transvaginal ultrasound (TVS), the first-line test, still suffers from limited specificity and operator dependence. We introduce a learned early-fusion (joint projection) hybrid that couples EfficientNet-B7 (local descriptors) with a Swin Transformer (hierarchical global context) to classify eight ovarian tumor categories from 2D TVS. Using the public, de-identified OTU-2D dataset (n = 1,469 images across eight histopathologic classes), we conducted patient-level, stratified 5-fold cross-validation repeated 10×. To address class imbalance while preventing leakage, training used train-only oversampling, ultrasound-aware augmentations, and strong regularization; validation/test folds were never resampled. The hybrid achieved AUC 0.9904, accuracy 92.13%, sensitivity 92.38%, and specificity 98.90%, outperforming single CNN or ViT baselines. A soft ensemble of the top hybrids further improved performance to AUC 0.991, accuracy 93.3%, sensitivity 93.6%, and specificity 99.0%. Beyond discrimination, we provide deployment-oriented evaluation: isotonic calibration yielded reliable probabilities, decision-curve analysis showed net clinical benefit across 5–20% risk thresholds, entropy-based uncertainty supported confidencebased triage, and Grad-CAM highlighted clinically salient regions. All metrics are reported with 95% bootstrap confidence intervals, and the evaluation protocol preserves real-world data distributions. Taken together, this work advances ovarian ultrasound AI from accuracy-only reporting to calibrated, explainable, and uncertainty-aware decision support, offering a reproducible reference framework for multiclass ovarian ultrasound and a clear path toward clinical integration and prospective validation. | es |
| dc.language.iso | en | es |
| dc.rights | Attribution-NonCommercial-NoDerivatives 4.0 Internacional | * |
| dc.rights | Attribution-NonCommercial-NoDerivatives 4.0 Internacional | * |
| dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/4.0/ | * |
| dc.subject | Ovarian cancer | es |
| dc.subject | Ultrasound imaging | es |
| dc.subject | Deep learning | es |
| dc.subject | CNN | es |
| dc.subject | Vision transformer | es |
| dc.subject | Hybrid model | es |
| dc.subject | Early diagnosis | es |
| dc.title | Early-fusion hybrid CNN-transformer models for multiclass ovarian tumor ultrasound classification | es |
| dc.type | journal article | es |
| dc.rights.accessRights | open access | es |
| dc.journal.title | Frontiers in Artificial Intelligence | es |
| dc.volume.number | 8 | es |
| dc.description.discipline | Ingeniería, Industria y Construcción | es |
| dc.identifier.doi | 10.3389/frai.2025.1679310 | es |
| dc.description.faculty | Escuela Politécnica | es |





