This paper presents a scalable object detection workflow for detecting objects, such as settlements, from remotely sensed (RS) imagery. We have successfully deployed this workflow on Titan supercomputer and utilized it for the task of mapping human settlement at a country scale.The performance of various stages in the workflow was analyzed before making it operational.The workflow implemented various strategies to address issues such as suboptimal resource utilization and long-tail effects due to unbalanced image workload, data loss due to runtime failures, and maximum wall-time constraints imposed by Titan's job scheduling policy. A mean shift clustering-based static load balancing strategy was implemented, which partitions the image load such that each partition contained similar-sized images. Furthermore, a checkpoint-restart strategy was added in the workflow as a fault-tolerance mechanism to prevent the data losses due to unforeseen runtime failures. The performance of the above-mentioned strategies was observed in various scenarios, such as node failure, exceeding wall time, and successful completion. Using this workflow, we have processed an RS data set that has a spatial resolution of 0.31 m and is comprised of 685 675 km 2 of area of the Republic of Zambia in under six hours using 5426 nodes of the Titan supercomputer. KEYWORDS convolutional neural network, deep learning, fault tolerance, HPC, human settlement mapping, load balancing 1 INTRODUCTION The recent development in sensor electronics is enabling remote sensing (RS) satellites to capture data at a submeter scale. This advancement in RS technology offers new opportunities to understand Earth's features at a finer spatial scale and motivates the development of novel approaches for complex problems such as mapping human settlements. However, as an effect of the development in sensor technology, a rapid growth in the cumulative volume of Earth observation (EO) archives has been observed during the last few years. 1 A typical very-high-resolution RS imagery contains billions of pixels, and for a country-scale mapping, thousands of such images need to be processed, which is computationally demanding. This computational load can pose problems during time-critical events, such as with flooding and wild fire natural disasters, when real-time/near-real-time mapping of human settlement in the affected area is necessary. However, high-performance computing (HPC)-driven approaches for processing such huge volume of data can mitigate this problem.