2015 IEEE International Conference on Multimedia &Amp; Expo Workshops (ICMEW) 2015
DOI: 10.1109/icmew.2015.7169816
|View full text |Cite
|
Sign up to set email alerts
|

Food image recognition using deep convolutional network with pre-training and fine-tuning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
121
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 267 publications
(128 citation statements)
references
References 9 publications
0
121
0
Order By: Relevance
“…They used the pre-trained AlexNet model as a feature extractor and integrated both CNN features and Fisher Vector encoded conventional SIFT and color features. Yanai et al [21] fine-tuned the AlexNet model and achieved the best results on public food datasets so far, with top-1 accuracy of 78.8% for UEC-FOOD-100 dataset and 67.6% for UEC-FOOD-256 [22] (another Japanese food image dataset with 256 classes). Their works showed that the recognition performance on small image datasets like UEC-FOOD-256 and UEC-FOOD-100 (both of which contained 100 images for each class) can be boosted by fine-tuning the CNN network which was pre-trained on a large dataset of similar objects.…”
Section: Food Image Recognitionmentioning
confidence: 99%
“…They used the pre-trained AlexNet model as a feature extractor and integrated both CNN features and Fisher Vector encoded conventional SIFT and color features. Yanai et al [21] fine-tuned the AlexNet model and achieved the best results on public food datasets so far, with top-1 accuracy of 78.8% for UEC-FOOD-100 dataset and 67.6% for UEC-FOOD-256 [22] (another Japanese food image dataset with 256 classes). Their works showed that the recognition performance on small image datasets like UEC-FOOD-256 and UEC-FOOD-100 (both of which contained 100 images for each class) can be boosted by fine-tuning the CNN network which was pre-trained on a large dataset of similar objects.…”
Section: Food Image Recognitionmentioning
confidence: 99%
“…Another of the topics studied is the evaluation of whether, for these tasks, it is better to train a network from scratch (full training) or to use fine tuning of a pre-trained network. There is a reference that analyzes this specific aspect in medical images [21] and something similar in images of food [34,35], but no bibliography in the field of architectural heritage is known.…”
Section: Image Classificationmentioning
confidence: 99%
“…In [17,18], the user is requested to draw a bounding box around the food items, while in [19], the user must mark the initial seeds before starting to grow segments. In this work, a semi-automatic method is designed which can be run on smartphones in a user-friendly manner.…”
Section: Segmentationmentioning
confidence: 99%