2020
DOI: 10.11591/ijece.v10i1.pp1017-1026
|View full text |Cite
|
Sign up to set email alerts
|

ResSeg: Residual encoder-decoder convolutional neural network for food segmentation

Abstract: This paper presents the implementation and evaluation of different convolutional neural network architectures focused on food segmentation. To perform this task, it is proposed the recognition of 6 categories, among which are the main food groups (protein, grains, fruit, and vegetables) and two additional groups, rice and drink or juice. In addition, to make the recognition more complex, it is decided to test the networks with food dishes already started, i.e. during different moments, from its serving to its … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 21 publications
0
6
0
Order By: Relevance
“…We employed SegNet [38], which was applied to the food extraction in Refs. [14,15], as the food extraction using low-training-cost DNN. SegNet is trained by UNIMIB2016 [39], which includes 1017 annotated food images available on the Internet.…”
Section: Precision = Tp/(tp + Fp)mentioning
confidence: 99%
See 2 more Smart Citations
“…We employed SegNet [38], which was applied to the food extraction in Refs. [14,15], as the food extraction using low-training-cost DNN. SegNet is trained by UNIMIB2016 [39], which includes 1017 annotated food images available on the Internet.…”
Section: Precision = Tp/(tp + Fp)mentioning
confidence: 99%
“…Refs. [12][13][14][15] proposed food extraction using DNN, requiring pixel-wise annotated food images as a training dataset. The food regions include various shapes, colors, and textures [12,16].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Image description is a process to extract the visual content of foods. In particular, local feature is more suitable to represent food features as the properties of local features that capture minuscule parts of the food beside its robustness towards illumination, scale, rotation, and orientation which made it capable to deal with the cluttered appearance of foods [1,5,6]. The interest points that were detected and described have produced a high volume and diverse features that require the features to be transformed into another more simplified representation by using certain feature encoding technique.…”
Section: Introductionmentioning
confidence: 99%
“…Besides CNN, other artificial intelligence methods have been developed for the classification of patterns, such as the fast R-CNN [15] and the DAG-CNN [16,17], in the first case is former a stage of extraction of a Regions of Interest (ROIs) that is responsible for of detecting desired elements in the input image, extracting them and entering them into a CNN for classification, as explained in [18], while the last consists of a branched structure where each branch contains a sequence of convolutional layers whose filters vary of dimension, to extract characteristics of greater and smaller size of the input image, and in the end to unify the results to give a classification, as indicated in [19].…”
Section: Introductionmentioning
confidence: 99%