2021
DOI: 10.1007/s00521-021-06129-w
|View full text |Cite
|
Sign up to set email alerts
|

Scalable multi-product inventory control with lead time constraints using reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(9 citation statements)
references
References 53 publications
0
7
0
Order By: Relevance
“…However, it must be assumed that the complex structure of current SCs, especially global ones with many stages and nodes, the number of variables included in the modeled problem and its intrinsically stochastic condition imply that the modeling of real cases with the reinforcement learning methodology, but without the additional assistance of other methods, constitutes a considerable challenge. Only through the gradual incorporation of the DRL methodology [69], a combination of the reinforcement learning methodology with deep learning-another ML methodology that uses artificial neural networks to transform a set of inputs into a set of outputs, that solve tasks that involve handling complex and high-dimensional raw input data sets [91]-has it been possible to begin to consider the study of SCs with certain complexity, e.g.,: (i) the multistage SC problem of Alves and Mateus [67], validated with a four-stage SC scenario and two nodes per stage, local inventories, lead time, a single product, and demand uncertainty; (ii) the capacitated SC problem of Peng et al [68], validated with a three-stage SC scenario, one node in the first, two in the second and three in the last stage, capacitated production, independent, stochastic and seasonal demand, and a single product; (iii) the case of Meisheri et al [92] who, despite restricting the validation of their retailers' inventory replenishment to the last SC layers, i.e., warehouse and retailer, considers the existence of product variety, with instances of 100 and 220 products-to substantially increase combinatorial computation-and incorporates lead time, limited storage capacity, cross-product restrictions, and weight and volume transportation restrictions. Computational limitations in this regard are manifested as the size of the problem to be solved in terms of the size of the input dataset, and especially the size of the modeled problem's observation space.…”
Section: Discussionmentioning
confidence: 99%
“…However, it must be assumed that the complex structure of current SCs, especially global ones with many stages and nodes, the number of variables included in the modeled problem and its intrinsically stochastic condition imply that the modeling of real cases with the reinforcement learning methodology, but without the additional assistance of other methods, constitutes a considerable challenge. Only through the gradual incorporation of the DRL methodology [69], a combination of the reinforcement learning methodology with deep learning-another ML methodology that uses artificial neural networks to transform a set of inputs into a set of outputs, that solve tasks that involve handling complex and high-dimensional raw input data sets [91]-has it been possible to begin to consider the study of SCs with certain complexity, e.g.,: (i) the multistage SC problem of Alves and Mateus [67], validated with a four-stage SC scenario and two nodes per stage, local inventories, lead time, a single product, and demand uncertainty; (ii) the capacitated SC problem of Peng et al [68], validated with a three-stage SC scenario, one node in the first, two in the second and three in the last stage, capacitated production, independent, stochastic and seasonal demand, and a single product; (iii) the case of Meisheri et al [92] who, despite restricting the validation of their retailers' inventory replenishment to the last SC layers, i.e., warehouse and retailer, considers the existence of product variety, with instances of 100 and 220 products-to substantially increase combinatorial computation-and incorporates lead time, limited storage capacity, cross-product restrictions, and weight and volume transportation restrictions. Computational limitations in this regard are manifested as the size of the problem to be solved in terms of the size of the input dataset, and especially the size of the modeled problem's observation space.…”
Section: Discussionmentioning
confidence: 99%
“…In the retail industry, having multiple products with uncertain demands and different lead times makes determining the optimal inventory replenishment policy highly challenging (Meisheri et al 2021). The authors of Meisheri et al (2021) addressed these challenges in a multi-period and multi-product system using DRL.…”
Section: Inventory Managementmentioning
confidence: 99%
“…Two DRL techniques applied in the reviewed papers are Deep Q-Network (DQN) proposed by Mnih et al (2015) and Proximal Policy Optimization (PPO) introduced by Schulman et al (2017). Meisheri et al (2021) employed both the DQN and PPO methods to determine the optimum replenishment decisions for retail businesses under uncertain demand, having multiple products with different lead times and cross-product constraints. Their results showed a better performance for the DQN.…”
Section: Deep Reinforcement Learning (Drl) and Its Applications In Th...mentioning
confidence: 99%
“…DRL methods were also considered for inventory control of multiple products. In order to drive such problem effectively, [16] utilized multi-agent reinforcement learning (MARL) method to effectively replenish the products without giving unfair treatment to certain products.…”
Section: B Reinforcement Learning In Inventory Managementmentioning
confidence: 99%