2023
DOI: 10.1016/j.jocs.2022.101913
|View full text |Cite
|
Sign up to set email alerts
|

Optimal checkpointing for adjoint multistage time-stepping schemes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…Griewank and Walther (2000) and Symes (2007) presented optimal checkpointing, which avoids excessive memory usage by balancing I/O and computational overhead optimally. This approach, which was initially developed for generic adjoint-state methods on CPUs (Griewank & Walther, 2000), has been used successfully in seismic imaging (Symes, 2007) and machine learning (Chen et al, 2016) and has recently been extended to multi-stage timestepping (Zhang & Constantinescu, 2022). By adding on-the-fly compression and decompression of the checkpointed forward wavefields, Kukreja et al (2020) further reduced the computational overhead of optimal checkpointing.…”
Section: Introductionmentioning
confidence: 99%
“…Griewank and Walther (2000) and Symes (2007) presented optimal checkpointing, which avoids excessive memory usage by balancing I/O and computational overhead optimally. This approach, which was initially developed for generic adjoint-state methods on CPUs (Griewank & Walther, 2000), has been used successfully in seismic imaging (Symes, 2007) and machine learning (Chen et al, 2016) and has recently been extended to multi-stage timestepping (Zhang & Constantinescu, 2022). By adding on-the-fly compression and decompression of the checkpointed forward wavefields, Kukreja et al (2020) further reduced the computational overhead of optimal checkpointing.…”
Section: Introductionmentioning
confidence: 99%
“…This allows traditional DRAM to be used as an extra cache level (Wu et al 2017, Stanzione et al 2020. The last mentioned approach, trading computation for storage, includes checkpointing methods (Griewank and Walther 1997, Wang et al 2009, Zhang and Constantinescu 2023 and sparse decomposition (Zhao et al 2020). More generally, this means storing the inputs required to regenerate the desired output rather than storing the output directly (Akturk and Karpuzcu 2018).…”
Section: Introductionmentioning
confidence: 99%
“…Griewank & Walther (2000) proposed a checkpointing algorithm, which is optimal under certain assumptions, including that the number of steps is known in advance, and that all the storage has equal access cost. Subsequent authors have produced checkpointing algorithms that relax these requirements in various ways, such as by accounting for different types of storage (e.g., memory and disk) or by not requiring the number of steps to be known in advance, for example Stumm & Walther (2009), Aupy et al (2016), Schanen et al (2016), Aupy & Herrmann (2017), Herrmann & Pallez (2020), James R. Maddison (2023), and Zhang & Constantinescu (2023).…”
mentioning
confidence: 99%