An analysis of layer-freezing strategies for enhanced transfer learning in YOLO architectures

Dobrzycki, Andrés Daniel ORCID: https://orcid.org/0009-0005-9756-4623, Bernardos Barbolla, Ana María ORCID: https://orcid.org/0000-0001-7678-8154 and Casar Corredera, José Ramón ORCID: https://orcid.org/0000-0003-3851-9038 (2025). An analysis of layer-freezing strategies for enhanced transfer learning in YOLO architectures. "Mathematics", v. 13 (n. 15); p. 2539. ISSN 2227-7390. https://doi.org/10.3390/math13152539.

Descripción

Título: An analysis of layer-freezing strategies for enhanced transfer learning in YOLO architectures
Autor/es:
Tipo de Documento: Artículo
Título de Revista/Publicación: Mathematics
Fecha: 7 Agosto 2025
ISSN: 2227-7390
Volumen: 13
Número: 15
Materias:
ODS:
Palabras Clave Informales: YOLO; layer freezing; transfer learning; fine-tuning; object detection
Escuela: E.T.S.I. Telecomunicación (UPM)
Departamento: Señales, Sistemas y Radiocomunicaciones
Licencias Creative Commons: Reconocimiento

Texto completo

[thumbnail of 10383666.pdf] PDF (Portable Document Format) - Se necesita un visor de ficheros PDF, como GSview, Xpdf o Adobe Acrobat Reader
Descargar (3MB)

Resumen

The You Only Look Once (YOLO) architecture is crucial for real-time object detection. However, deploying it in resource-constrained environments such as unmanned aerial vehicles (UAVs) requires efficient transfer learning. Although layer freezing is a common technique, the specific impact of various freezing configurations on contemporary YOLOv8 and YOLOv10 architectures remains unexplored, particularly with regard to the interplay between freezing depth, dataset characteristics, and training dynamics. This research addresses this gap by presenting a detailed analysis of layer-freezing strategies. We systematically investigate multiple freezing configurations across YOLOv8 and YOLOv10 variants using four challenging datasets that represent critical infrastructure monitoring. Our methodology integrates a gradient behavior analysis (L2 norm) and visual explanations (Grad-CAM) to provide deeper insights into training dynamics under different freezing strategies. Our results reveal that there is no universal optimal freezing strategy but, rather, one that depends on the properties of the data. For example, freezing the backbone is effective for preserving general-purpose features, while a shallower freeze is better suited to handling extreme class imbalance. These configurations reduce graphics processing unit (GPU) memory consumption by up to 28% compared to full fine-tuning and, in some cases, achieve mean average precision (mAP@50) scores that surpass those of full fine-tuning. Gradient analysis corroborates these findings, showing distinct convergence patterns for moderately frozen models. Ultimately, this work provides empirical findings and practical guidelines for selecting freezing strategies. It offers a practical, evidence-based approach to balanced transfer learning for object detection in scenarios with limited resources.

Proyectos asociados

Tipo
Código
Acrónimo
Responsable
Título
Horizonte Europa
01103386
Sin especificar
Sin especificar
Sin especificar
Gobierno de España
MIA.2021.M04.0008
IA4TES
Sin especificar
Tecnologías Inteligentes Avanzadas para la transición Energética Sostenible
Gobierno de España
PID2023-151605OB-C21
ASPID
Sin especificar
Avances en Sistemas de Protección e Inteligencia con enjambres de Drones

Más información

ID de Registro: 95361
Identificador DC: https://oa.upm.es/95361/
Identificador OAI: oai:oa.upm.es:95361
URL Portal Científico: https://portalcientifico.upm.es/es/ipublic/item/10383666
Identificador DOI: 10.3390/math13152539
URL Oficial: https://www.mdpi.com/2227-7390/13/15/2539
Depositado por: iMarina Portal Científico
Depositado el: 16 Abr 2026 12:27
Ultima Modificación: 16 Abr 2026 12:27