2018
DOI: 10.1016/j.procs.2018.07.112
|View full text |Cite
|
Sign up to set email alerts
|

YOLO based Human Action Recognition and Localization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
41
0
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 147 publications
(62 citation statements)
references
References 4 publications
0
41
0
1
Order By: Relevance
“…2 CNN: Image features are extracted by convolution operation, according to color, edge, texture and so on [16].…”
Section: Baseline Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…2 CNN: Image features are extracted by convolution operation, according to color, edge, texture and so on [16].…”
Section: Baseline Methodsmentioning
confidence: 99%
“…It uses the sum of the squared errors (SSE) as the loss function [15]. The YOLO algorithm divides images into S*S grids, and the output of each grid is (B*5+C) dimensions, which include the location information, the confidence of the border box [16], and the number of categories. However, these factors have different effects on the accuracy of the object recognition of vehicles.…”
Section: Loss Functionmentioning
confidence: 99%
“…Several types of neural networks exist and the YOLO is the one chosen for this project, for the following reasons [17,18]:…”
Section: Neural Networkmentioning
confidence: 99%
“…Neural Network training A first training has been carried out using a tiny YOLOv2 pretrained with the COCO dataset. YOLO requires some files to start training which are [17,18] a configuration file with all layers of YOLO architecture (described in Figure 2). • pre-trained convolutional weights.…”
Section: 42mentioning
confidence: 99%
“…It uses a separate CNN model to realize end-to-end target detection, divides the input images into 7 × 7 grids, and then each cell is responsible for predicting the targets in which the center points fall in the grid; when the pumping unit or head working fall in some grid, this grid is responsible for predicting them, compares the predicted value with the real value, and calculates the predicted loss. The core idea is to directly manipulate the whole picture by inputting a figure directly in the output layer for each grid to predict the B bounding box location information and the confidence score of the bounding box [26].…”
Section: Using Yolov3 As a Detector Of The Ylts Framework To Detect Tmentioning
confidence: 99%