Single image depth estimation works fail to separate foreground elements because they can easily be confounded with the background. To alleviate this problem, we propose the use of a semantic segmentation procedure that adds information to a depth estimator, in this case, a 3D Convolutional Neural Network (CNN)—segmentation is coded as one-hot planes representing categories of objects. We explore 2D and 3D models. Particularly, we propose a hybrid 2D–3D CNN architecture capable of obtaining semantic segmentation and depth estimation at the same time. We tested our procedure on the SYNTHIA-AL dataset and obtained σ3=0.95, which is an improvement of 0.14 points (compared with the state of the art of σ3=0.81) by using manual segmentation, and σ3=0.89 using automatic semantic segmentation, proving that depth estimation is improved when the shape and position of objects in a scene are known.
In the medical area, the detection of invasive ductal carcinoma is the most common sub-type of all breast cancers; about 80% of all breast cancers are invasive ductal carcinomas. Detection of this type of cancer shows a great challenge for specialist doctors since the digital images of the sample must be analyzed by sections because the spatial dimensions of this kind of image are above 50k × 50k pixels; doing this operation manually takes long time to determine if the patient suffers this type of cancer. Time is essential for the patient because this cancer can invade quickly other parts of the body. Its name reaffirms this characteristic, with the term "invasive" forming part of its name. With the purpose of solving this task, we propose an automatic methodology consisting in improving the performance of a convolutional neural network that classifies images containing invasive ductal carcinoma cells by highlighting cancer cells using several preprocessing methods such as histogram stretching and contrast enhancement. In this way, characteristics of the sub-images are extracted from the panoramic sample and it is possible to learn to classify them in a better way.
Glaucoma detection is an important task, as this disease can affect the optic nerve, and this could lead to blindness. This can be prevented with early diagnosis, periodic controls, and treatment so that it can be stopped and prevent visual loss. Usually, the detection of glaucoma is carried out through various examinations such as tonometry, gonioscopy, pachymetry, etc. In this work, we carry out this detection by using images obtained through retinal cameras, in which we can observe the state of the optic nerve. This work addresses an accurate diagnostic methodology based on Convolutional Neural Networks (CNNs) to classify these optical images. Most works require a large number of images to train their CNN architectures, and most of them use the whole image to perform the classification. We will use a small dataset containing 366 examples to train the proposed CNN architecture and we will only focus on the analysis of the optic disc by extracting it from the full image, as this is the element that provides the most information about glaucoma. We experiment with different RGB channels and their combinations from the optic disc, and additionally, we extract depth information. We obtain accuracy values of 0.945, by using the GB and the full RGB combination, and 0.934 for the grayscale transformation. Depth information did not help, as it limited the best accuracy value to 0.934.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.