In this study, the computational development conducted was based on Convolutional Neural Networks (CNNs), and the You Only Look Once (YOLO) algorithm to detect vehicles from aerial images and calculate the safe distance between them. We analyzed a dataset composed of 896 images, recorded in videos by a DJI Spark Drone. The training set used 60% of the images, 20% for validation, and 20% for the tests. Tests were performed to detect vehicles in different configurations, and the best result was achieved using the YOLO Full-608, with a mean Average Precision(mAP) of 95.6%. The accuracy of the results encourages the development of systems capable of estimating the safe distance between vehicles in motion, allowing mainly to minimize the risk of accidents.
This chapter intends to present the main techniques for detecting objects within images. In recent years there have been remarkable advances in areas such as machine learning and pattern recognition, both using convolutional neural networks (CNNs). It is mainly due to the increased parallel processing power provided by graphics processing units (GPUs). In this chapter, the reader will understand the details of the state-of-the-art algorithms for object detection in images, namely, faster region convolutional neural network (Faster RCNN), you only look once (YOLO), and single shot multibox detector (SSD). We will present the advantages and disadvantages of each technique from a series of comparative tests. For this, we will use metrics such as accuracy, training difficulty, and characteristics to implement the algorithms. In this chapter, we intend to contribute to a better understanding of the state of the art in machine learning and convolutional networks for solving problems involving computational vision and object detection.
Chess is one of the most researched domains in the annals of artificial intelligence. The main objective of this research is to develop a platform that can determine piece positioning during chess games. Digital image processing methods and real-time object detection (YOLO version 4) algorithms were used during computational development. The problem entails analyzing images captured during a chess game and determining the location of each square on the board, as well as the position of each piece in play. This procedure is repeated at each game turn, enabling the developed system to save and watch all piece moves during a game. The obtained results demonstrate the system’s reliability and feasibility.
Traffic accidents are among the most worrying problems in modern life, often caused by human operational errors such as inattention, distraction, and misbehavior. Vehicle speed detection and safety distance measurement can help reduce these accidents. In this study, the computational development conducted was based on Convolutional Neural Networks (CNNs) and the You Only Look Once (YOLO) algorithm to detect vehicles from aerial images and calculate the safe distance and the vehicle’s speed on Brazilian highways. The investigation was conducted to model the YOLO algorithm for detecting vehicles in different network architecture configurations. The best results were obtained with the YOLO Full-608, reaching a mean Average Precision (mAP) of 97.44%. Additional computer vision approaches have been developed to calculate the speed of the moving vehicle and the safe distance between them. Therefore, the developed system allows that, based on detecting the safe distance between moving vehicles on the highways, accidents are predicted and possibly avoided.
In this paper, state-of-the-art architectures of Convolutional Neural Networks (CNNs) are explained and compared concerning authorship classification of famous paintings. The chosen CNNs architectures were VGG-16, VGG-19, Residual Neural Networks (ResNet), and Xception. The used dataset is available on the website Kaggle, under the title “Best Artworks of All Time”. Weighted classes for each artist with more than 200 paintings present in the dataset were created to represent and classify each artist’s style. The performed experiments resulted in an accuracy of up to 95% for the Xception architecture with an average F1-score of 0.87, 92% of accuracy with an average F1-score of 0.83 for the ResNet in its 50-layer configuration, while both of the VGG architectures did not present satisfactory results for the same amount of epochs, achieving at most 60% of accuracy.
The computational tool developed in this study is based on convolutional neural networks and the You Only Look Once (YOLO) algorithm for detecting and tracking mice in videos recorded during behavioral neuroscience experiments. We analyzed a set of data composed of 13622 images, made up of behavioral videos of three important researches in this area. The training set used 50% of the images, 25% for validation, and 25% for the tests. The results show that the mean Average Precision (mAP) reached by the developed system was 90.79% and 90.75% for the Full and Tiny versions of YOLO, respectively. Considering the high accuracy of the results, the developed work allows the experimentalists to perform mice tracking in a reliable and non-evasive way.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.