2022
DOI: 10.1109/jiot.2022.3166110
|View full text |Cite
|
Sign up to set email alerts
|

D3PG: Dirichlet DDPG for Task Partitioning and Offloading With Constrained Hybrid Action Space in Mobile-Edge Computing

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(9 citation statements)
references
References 41 publications
0
7
0
Order By: Relevance
“…Within the realm of DRL, DDPG stands out for its effectiveness in handling continuous action spaces, attributed to its utilization of an actor-critic architecture that facilitates stable learning [10], [47]. In our problem (21), the action space comprises both continuous and discrete variables, culminating in a hybrid action space [48]. To navigate this complexity, one could manipulate the continuous action space to select discrete actions [48], [49], e.g., by generating a 10-dimensional action vector with values ranging between 0 and 1, and subsequently choosing the action corresponding to the position of the maximum value.…”
Section: Gdm-based Ddpg With Llms Interaction For Joint Resource Allo...mentioning
confidence: 99%
See 1 more Smart Citation
“…Within the realm of DRL, DDPG stands out for its effectiveness in handling continuous action spaces, attributed to its utilization of an actor-critic architecture that facilitates stable learning [10], [47]. In our problem (21), the action space comprises both continuous and discrete variables, culminating in a hybrid action space [48]. To navigate this complexity, one could manipulate the continuous action space to select discrete actions [48], [49], e.g., by generating a 10-dimensional action vector with values ranging between 0 and 1, and subsequently choosing the action corresponding to the position of the maximum value.…”
Section: Gdm-based Ddpg With Llms Interaction For Joint Resource Allo...mentioning
confidence: 99%
“…In our problem (21), the action space comprises both continuous and discrete variables, culminating in a hybrid action space [48]. To navigate this complexity, one could manipulate the continuous action space to select discrete actions [48], [49], e.g., by generating a 10-dimensional action vector with values ranging between 0 and 1, and subsequently choosing the action corresponding to the position of the maximum value. This methodology is plausible due to the continuous relaxation of discrete action spaces, allowing gradient-based optimization methods, like DDPG, to operate.…”
Section: Gdm-based Ddpg With Llms Interaction For Joint Resource Allo...mentioning
confidence: 99%
“…Reference 20 proposes a low‐complexity algorithm based on queuing theory in MEC networks, which minimizes the energy consumption and delay of mobile terminals by optimizing spectrum resources and edge server computing resources. In reference 21, an equal power allocation scheme for uplink is proposed, based on which the energy consumption of equipment is optimized.…”
Section: Related Workmentioning
confidence: 99%
“…However, the immediate reward was also affected by other actions; thus, the reward could not be maximized. L. Ale et al [37] proposed the Dirichlet deep deterministic policy (D3PG) to minimize the service times of tasks. They improved the performance by adding noise to the action, based on the Ornstein-Uhlenbeck process in the DDPG structure.…”
Section: Rl-based Computation Offloadingmentioning
confidence: 99%
“…TADPG [34] Single-agent B. Gu [35] Multi-agent DC-DRL [7] Multi-agent L. Ale et al [37] ding53 Single-agent A. Oadder et al [38] Single-agent Proposed Multi-agent…”
Section: Scheme Temporal State Experience Priority Agentmentioning
confidence: 99%