2014
DOI: 10.1007/978-3-319-05083-6_5
|View full text |Cite
|
Sign up to set email alerts
|

One-Shot Approaches to Design Optimzation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
6
2

Relationship

3
5

Authors

Journals

citations
Cited by 25 publications
(19 citation statements)
references
References 42 publications
0
19
0
Order By: Relevance
“…The MGRIT algorithm replaces the O(N ) sequential time stepping algorithm with a highly parallel O(N ) multigrid algorithm [56]. Essentially, a sequence of coarser temporal grids are used to accelerate the solution of the fine grid (5). The MGRIT algorithm easily extends to nonlinear problems with the Full Approximation Storage (FAS) nonlinear multigrid method [6], which is the same nonlinear multigrid cycling used by PFASST [10].…”
Section: Multigrid Reduction In Time Using Xbraidmentioning
confidence: 99%
“…The MGRIT algorithm replaces the O(N ) sequential time stepping algorithm with a highly parallel O(N ) multigrid algorithm [56]. Essentially, a sequence of coarser temporal grids are used to accelerate the solution of the fine grid (5). The MGRIT algorithm easily extends to nonlinear problems with the Full Approximation Storage (FAS) nonlinear multigrid method [6], which is the same nonlinear multigrid cycling used by PFASST [10].…”
Section: Multigrid Reduction In Time Using Xbraidmentioning
confidence: 99%
“…In the preceding section, we showed that there exist stepsizes α G and (projected Newton) preconditioners C that guarantee that the extended mapping G satisfies the contraction condition (2) and, thus, allow convergence of the overall method. A necessary condition for their existence was that there is only a slight coupling of the variables by the constraints, namely, cG or dG are sufficiently small.…”
Section: Enforcing Contraction For the General Casementioning
confidence: 97%
“…• Hessian approximation: In order to prove convergence of the simultaneous One-shot method on a theoretical level, the preconditioners B θ , B W , B µ should approximate the Hessian of an augmented Lagrangian function that involves the residual of the state and adjoint equations (see [9] and references therein). Numerically, we approximate the Hessian through consecutive limited-memory BFGS updates based on the current reduced gradient (thus assuming that the residual term is small).…”
Section: Algorithm 2 Simultaneous Layer-parallel Trainingmentioning
confidence: 99%
“…They aim at solving the optimization problem in an all-at-once fashion, updating the optimization parameters simultaneously while solving for the time-dependent system state. Here, we apply the One-shot method [9,32] to solve the training problem simultaneously for the network state and parameters. In this approach, network parameter updates are based on inexact gradient information resulting from early stopping of the layer-parallel multigrid iteration.…”
Section: Introductionmentioning
confidence: 99%