TY - JOUR
T1 - A Survey on Batch Training in Genetic Programming
AU - Rosenfeld, Liah
AU - Vanneschi, Leonardo
N1 - info:eu-repo/grantAgreement/FCT/6817 - DCRRNI ID/UIDB%2F04152%2F2020/PT#
https://doi.org/10.54499/UIDB/04152/2020#
Rosenfeld, L., & Vanneschi, L. (2025). A Survey on Batch Training in Genetic Programming. Genetic Programming And Evolvable Machines, 26, 1-28. Article 2. https://doi.org/10.1007/s10710-024-09501-6 --- Open access funding provided by FCT|FCCN (b-on). This work was supported by national funds through FCT (Fundação para a Ciência e a Tecnologia), under the project - UIDB/04152/2020 - Centro de Investigação em Gestão de Informação (MagIC)/NOVA IMS (https://doi.org/10.54499/UIDB/04152/2020).
PY - 2025/6
Y1 - 2025/6
N2 - In Machine Learning (ML), the use of subsets of training data, referred to as batches, rather than the entire dataset, has been extensively researched to reduce computational costs, improve model efficiency, and enhance algorithm generalization. Despite extensive research, a clear definition and consensus on what constitutes batch training have yet to be reached, leading to a fragmented body of literature that could otherwise be seen as different facets of a unified methodology. To address this gap, we propose a theoretical redefinition of batch training, creating a clearer and broader overview that integrates diverse perspectives. We then apply this refined concepjavascript:void(0);t specifically to Genetic Programming (GP). Although batch training techniques have been explored in GP, the term itself is seldom used, resulting in ambiguity regarding its application in this area. This review seeks to clarify the existing literature on batch training by presenting a new and practical classification system, which we further explore within the specific context of GP. We also investigate the use of dynamic batch sizes in ML, emphasizing the relatively limited research on dynamic or adaptive batch sizes in GP compared to other ML algorithms. By bringing greater coherence to previously disjointed research efforts, we aim to foster further scientific exploration and development. Our work highlights key considerations for researchers designing batch training applications in GP and offers an in-depth discussion of future research directions, challenges, and opportunities for advancement.
AB - In Machine Learning (ML), the use of subsets of training data, referred to as batches, rather than the entire dataset, has been extensively researched to reduce computational costs, improve model efficiency, and enhance algorithm generalization. Despite extensive research, a clear definition and consensus on what constitutes batch training have yet to be reached, leading to a fragmented body of literature that could otherwise be seen as different facets of a unified methodology. To address this gap, we propose a theoretical redefinition of batch training, creating a clearer and broader overview that integrates diverse perspectives. We then apply this refined concepjavascript:void(0);t specifically to Genetic Programming (GP). Although batch training techniques have been explored in GP, the term itself is seldom used, resulting in ambiguity regarding its application in this area. This review seeks to clarify the existing literature on batch training by presenting a new and practical classification system, which we further explore within the specific context of GP. We also investigate the use of dynamic batch sizes in ML, emphasizing the relatively limited research on dynamic or adaptive batch sizes in GP compared to other ML algorithms. By bringing greater coherence to previously disjointed research efforts, we aim to foster further scientific exploration and development. Our work highlights key considerations for researchers designing batch training applications in GP and offers an in-depth discussion of future research directions, challenges, and opportunities for advancement.
KW - Genetic programming
KW - Batch training
KW - Sampling methods
KW - Generalization
KW - Overfitting
UR - https://www.webofscience.com/wos/woscc/full-record/WOS:001366871900001
UR - http://www.scopus.com/inward/record.url?scp=85211687955&partnerID=8YFLogxK
U2 - 10.1007/s10710-024-09501-6
DO - 10.1007/s10710-024-09501-6
M3 - Review article
SN - 1389-2576
VL - 26
SP - 1
EP - 28
JO - Genetic Programming And Evolvable Machines
JF - Genetic Programming And Evolvable Machines
M1 - 2
ER -