2020 IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI) 2020
DOI: 10.1109/ssiai49293.2020.9094606
|View full text |Cite
|
Sign up to set email alerts
|

Weeping and Gnashing of Teeth: Teaching Deep Learning in Image and Video Processing Classes

Abstract: In this rather informal paper and talk I will discuss my own experiences, feelings and evolution as an Image Processing and Digital Video educator trying to navigate the Deep Learning revolution. I will discuss my own ups and downs of trying to deal with extremely rapid technological changes, and how I have reacted to, and dealt with consequent dramatic changes in the relevance of the topics I've taught for three decades. I have arranged the discussion in terms of the stages, over time, of my progression deali… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 22 publications
0
2
0
Order By: Relevance
“…Researches showed that transferring existing classification networks to regression tasks performed not well [18,20,22]. As noted by A. C. Bovik [23], "Unlike human participation in crowdsourced picture labeling experiments like ImageNet, where each human label might need only 0.5-1.0 seconds to apply, human quality judgments on pictures generally required 10-20x that amount to time for a subject to feel comfortable in making their assessments on a Likert scale [24]." In general, a clip of video has a long duration including hundreds of images, and its perceptual quality difference from other videos with different content is extremely subtle.…”
Section: Introductionmentioning
confidence: 99%
“…Researches showed that transferring existing classification networks to regression tasks performed not well [18,20,22]. As noted by A. C. Bovik [23], "Unlike human participation in crowdsourced picture labeling experiments like ImageNet, where each human label might need only 0.5-1.0 seconds to apply, human quality judgments on pictures generally required 10-20x that amount to time for a subject to feel comfortable in making their assessments on a Likert scale [24]." In general, a clip of video has a long duration including hundreds of images, and its perceptual quality difference from other videos with different content is extremely subtle.…”
Section: Introductionmentioning
confidence: 99%
“…Researches showed that transferring existing classification networks to regression tasks performed not well [8], [3], [32]. As noted by A. C. Bovik [4], "Unlike human participation in crowdsourced picture labeling experiments like ImageNet, where each human label might need only 0.5-1.0 seconds to apply, human quality judgments on pictures generally required 10-20x that amount to time for a subject to feel comfortable in making their assessments on a Likert scale [10]." In general, a clip of video has a long duration including hundreds of images, and its perceptual quality difference from other videos with different content is extremely subtle.…”
Section: Introductionmentioning
confidence: 99%