2021
DOI: 10.1016/j.asoc.2021.107084
|View full text |Cite
|
Sign up to set email alerts
|

Guided parallelized stochastic gradient descent for delay compensation

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
16
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
9
1

Relationship

2
8

Authors

Journals

citations
Cited by 16 publications
(16 citation statements)
references
References 15 publications
0
16
0
Order By: Relevance
“…S j ( y i , x , i ) is defined as the state feature function corresponding to the marked position I in the observation sequence and it is applied to describe the influence of the observation sequence on the marked variable, as shown in Equation (4); λ j and μ k are parameters, and Z refers to the normalization factor to ensure the probability is correctly defined. The values of parameters λ j and μ k are estimated based on the stochastic gradient descent algorithm in this study because it has a faster convergence rate than other gradient descent algorithms (Sharma, 2021). tj)(yi+1,x,igoodbreak={11emif0.12emyi+1goodbreak=ya,yigoodbreak=yb0.12emand0.24emxigoodbreak=xa01emotherwise sj)(yi,x,igoodbreak={11emif0.12emyigoodbreak=yb0.12emand0.24emxigoodbreak=xa01emotherwise The details for POI name extraction from Chinese texts based on the improved CRF algorithm in this study mainly include sequence annotation, feature selection, feature annotation, template design, and model training and testing.…”
Section: Methodsmentioning
confidence: 99%
“…S j ( y i , x , i ) is defined as the state feature function corresponding to the marked position I in the observation sequence and it is applied to describe the influence of the observation sequence on the marked variable, as shown in Equation (4); λ j and μ k are parameters, and Z refers to the normalization factor to ensure the probability is correctly defined. The values of parameters λ j and μ k are estimated based on the stochastic gradient descent algorithm in this study because it has a faster convergence rate than other gradient descent algorithms (Sharma, 2021). tj)(yi+1,x,igoodbreak={11emif0.12emyi+1goodbreak=ya,yigoodbreak=yb0.12emand0.24emxigoodbreak=xa01emotherwise sj)(yi,x,igoodbreak={11emif0.12emyigoodbreak=yb0.12emand0.24emxigoodbreak=xa01emotherwise The details for POI name extraction from Chinese texts based on the improved CRF algorithm in this study mainly include sequence annotation, feature selection, feature annotation, template design, and model training and testing.…”
Section: Methodsmentioning
confidence: 99%
“…The time-complexity of GAN can be roughly given as O(nT Ld 2 ) where the new parameters L and T are layer-size and total iterations for a GAN. Its convergence rate with the Stochastic Gradient Descent would be O( 1T + σ 2 ) where σ 2 is the variance of the dataset [59]. See Section III for further details.…”
Section: B Generative Adversarial Network (Gan)mentioning
confidence: 99%
“…A high value has the risk of missing the lowest point of the slope. The GD algorithm performs two main steps: gradient computation and update of coefficients ( m , b ), which starts with the initialization of α, initial vector of [ m , b ] and the total number of iterations, then the algorithm stops when the termination criterion is meet, or just the overfitting starts [ 51 , 52 ]. In this paper, the GD algorithm stops when the shadowing achieves a mean value of 0 and random behavior.…”
Section: Post-processingmentioning
confidence: 99%