The intersection of deep learning and programmable logic controllers (PLCs) can lead to innovative applications in automation. One of the exciting application areas are gesture-based control systems for Automated Guided Vehicles (AGVs). AGVs are used in various industries for material handling, logistics, warehouse automation, etc. Traditionally, these vehicles are controlled using predefined routes or remote controls, but with gesture-based control, operators can communicate more naturally and efficiently. The incorporation of YOLO-Pose in YOLO versions 7 and 8 has elevated the YOLO algorithm to a leading tool for creating gesture recognition models. The YOLO algorithm employs convolutional neural networks (CNN) to detect objects in real-time. These latest YOLO models offer significantly improved accuracy, speed, and reduced training times. This paper presents the comparative results of 2D gesture recognition transfer learning models created using the YOLO v5, v7, and v8 models, along with the steps taken to implement the model in a PLCcontrolled AGV. Over 14,000 images were collected to build the models. A semi-automated approach was used to annotate them. Five models were created: two Keypoint models and three object detection models using transfer learning techniques with the same hyperparameters.