Pre-trained Encoders for Global Child Development: Transfer Learning Enables Deployment in Data-Scarce Settings
MMM Fahim, MR Karim
arXiv preprint · 2026 · preprint · arXiv:2601.20987
TL;DR
Pre-trained on 357K children across 44 countries, this encoder solves the cold-start problem: with only 50 samples it beats gradient boosting by 8–12%. Zero-shot to unseen countries still reaches AUC 0.84.
Abstract
A large number of children experience preventable developmental delays each year, yet deployment of machine learning in new countries is stymied by a data bottleneck: reliable models require thousands of samples, while new programs begin with fewer than 100. We introduce the first pre-trained encoder for global child development, trained on 357,709 children across 44 countries using UNICEF survey data. With only 50 training samples, the pre-trained encoder achieves an average AUC of 0.65 (95% CI: 0.56–0.72), outperforming cold-start gradient boosting by 8–12% across regions. At N = 500, the encoder achieves AUC of 0.73. Zero-shot deployment to unseen countries achieves AUCs up to 0.84. We apply a transfer learning bound to explain why pre-training diversity enables few-shot generalization, establishing that pre-trained encoders can transform the feasibility of ML for SDG 4.2.1 monitoring in resource-constrained settings.
BibTeX
@article{fahim2026pretrained,
title = {Pre-trained Encoders for Global Child Development: Transfer Learning Enables Deployment in Data-Scarce Settings},
author = {MMM Fahim and MR Karim},
year = {2026},
journal = {arXiv preprint},
eprint = {2601.20987},
archivePrefix = {arXiv},
url = {https://arxiv.org/abs/2601.20987},
}