2021
DOI: 10.3390/rs13122288
|View full text |Cite
|
Sign up to set email alerts
|

Two-Stream Dense Feature Fusion Network Based on RGB-D Data for the Real-Time Prediction of Weed Aboveground Fresh Weight in a Field Environment

Abstract: The aboveground fresh weight of weeds is an important indicator that reflects their biomass and physiological activity and directly affects the criteria for determining the amount of herbicides to apply. In precision agriculture, the development of models that can accurately locate weeds and predict their fresh weight can provide visual support for accurate, variable herbicide application in real time. In this work, we develop a two-stream dense feature fusion convolutional network model based on RGB-D data fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 17 publications
(13 citation statements)
references
References 60 publications
0
10
0
Order By: Relevance
“…For the depth images, the crop area was decreased to 256 × 256 pixels to minimize the effect of the background. The k-nearest neighbor method was applied to fill the pixels for which the depth image data were not acquired, and the average value of nine pixels was used [ 27 , 40 ].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…For the depth images, the crop area was decreased to 256 × 256 pixels to minimize the effect of the background. The k-nearest neighbor method was applied to fill the pixels for which the depth image data were not acquired, and the average value of nine pixels was used [ 27 , 40 ].…”
Section: Methodsmentioning
confidence: 99%
“…In addition, 3D data, such as RGB-D images, may provide more information on vertical changes in leaves and stems [ 23 , 24 , 25 , 26 ]. Quan et al developed a novel two-stream CNN model based on RGB-D images to estimate the fresh weight of aboveground weeds in a field on a high-end graphics processing unit (GPU) (2080Ti, NVIDIA, Santa Clara, CA, USA) [ 27 ]. The two-stream CNN model used a multi-input single-output (MISO) structure and dense network in the network blocks.…”
Section: Introductionmentioning
confidence: 99%
“…For example, long short-term memory (LSTM) and CNN were both used to extract features from social sensing signature data, which were concatenated with features extracted with CNN from spectral images to classify urban region functions [171]. The feature fusion may also be performed in intermediate layers in addition to the last layer [172][173][174][175], which is indicated by the dashed arrow lines in Figure 4b. In addition to concatenation, features can also be fused by maximum extraction operation, i.e., for each position in the feature vector, selecting the maximum values among the features extracted across all the data sources [175].…”
Section: Feature-level Fusionmentioning
confidence: 99%
“…Many studies have been conducted on body tracking and motion analysis using depth images [20][21][22][23]. There are two methods for creating depth images: extracting features from two-dimensional images and inferring depth through learning [24][25][26][27] or shooting with a 3D depth camera [28][29][30]. The former method has disadvantages in that an additional process is required to extract and learn features of an image, it takes a lot of time, and the accuracy is low.…”
Section: Motion Capture Systemmentioning
confidence: 99%