[go: up one dir, main page]

0% found this document useful (0 votes)
6 views3 pages

Assign 5

The document discusses various aspects of the CERAI algorithm and DDPG-based resource allocation, highlighting the roles of the experience replay pool, Actor and Critic networks, and cost calculations in cloud-edge computing. It explains that the experience replay pool is essential for sampling state transition tuples for training, while the Actor selects actions based on current states with added noise. Additionally, it outlines the main goal of resource allocation algorithms as minimizing long-term system costs and differentiates between PAMDP and MDP in terms of parameterized actions.

Uploaded by

Piyush Rawat
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
6 views3 pages

Assign 5

The document discusses various aspects of the CERAI algorithm and DDPG-based resource allocation, highlighting the roles of the experience replay pool, Actor and Critic networks, and cost calculations in cloud-edge computing. It explains that the experience replay pool is essential for sampling state transition tuples for training, while the Actor selects actions based on current states with added noise. Additionally, it outlines the main goal of resource allocation algorithms as minimizing long-term system costs and differentiates between PAMDP and MDP in terms of parameterized actions.

Uploaded by

Piyush Rawat
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

1. What is the primary role of the experience replay pool in the CERAI algorithm?

a. To store completed tasks for analysis after training.


b. To directly update the Actor and Critic networks after each action.
c. To store state transition tuples for sampling during gradient descent.
d. To track resource allocation history across multiple iterations.
Answer: c
Explanation: The experience replay pool stores state transition tuples that are later
sampled for gradient descent updates, which helps break correlations in training
data.
2. In the DDPG-based resource allocation algorithm, what action does the Actor main
network perform?
a. It selects an action based on the state and random noise.
b. It calculates the reward for the Critic network.
c. It directly updates the allocation record H.
d. It computes the next state for the edge node.
Answer: a

Explanation: The Actor main network uses the current state plus added
random noise to select an action for exploration.

3. How is the cost of collaborative cloud-edge computing calculated in a public cloud


environment?
a. Based solely on the on-demand instance cost.
b. By considering only the computing cost of cloud nodes.
c. By adding the cost of cloud instances (on-demand, reserved, and spot) and the
edge node.
d. By averaging the costs of edge and cloud nodes.

Answer: c

Explanation: In a public cloud setup, the total cost is calculated by summing the
costs of various cloud instance types (on-demand, reserved, spot) together with
the edge node’s cost.

4. What is the role of the Critic network in the Deep Deterministic Policy Gradient (DDPG)
algorithm?
a. To directly perform actions based on the policy.
b. To generate resource allocation policies independently.
c. To store experience in the replay pool.
d. To evaluate the Actor’s performance using a value function.

Answer: d
Explanation: The Critic network evaluates the Actor's performance by estimating
the value function, which guides the Actor’s policy updates.

5. What is the main goal of the resource allocation algorithms in cloud-edge computing?
a. To maximize the number of VMs allocated
b. To minimize the long-term cost of the system
c. To increase the computing time duration
d. To maximize the reward function
Answer: b
Explanation: The goal is to minimize the long-term cost of the system over the T
time slots by minimizing the sum of the costs over all time slots.

6. What are Availability Zones in AWS?


a. Geographic areas where AWS services are available
b. Multiple isolated locations/data centers within a region
c. Edge locations to deliver content to end users
d. Virtual networks defined by customers
Answer: b
Explanation: Availability Zones are defined as multiple isolated locations/data centers
within a region.

7. What is the main difference between PAMDP and MDP?


a. PAMDP has a different reward function
b. PAMDP uses a finite set of parameterized actions
c. PAMDP doesn't use neural networks
d. PAMDP is only used for private cloud environments
Answer: b
Explanation: The difference with the Markov decision process is that A is the finite set
of parameterized actions in PAMDP.

8. What does the Deep Deterministic Policy Gradient (DDPG) algorithm involve?
a. Only Actor networks.
b. Only Critic networks.
c. Both Actor and Critic networks.
d. Neither Actor nor Critic networks.
Answer: c
Explanation: DDPG involves both Actor and Critic networks to guide the
decision-making process in resource allocation.

9. What is the purpose of the experience replay pool in DDPG?


a. To store allocation records.
b. To sample experiences for training the networks.
c. To manage cloud costs.
d. To predict user demand.
Answer: b)
Explanation: The experience replay pool is used to sample experiences for updating
the Actor and Critic networks during training.

10. What is the Markov Decision Process (MDP) used for in resource allocation?
a. To model sequential decision-making problems.
b. To predict user demand.
c. To manage cloud costs.
d. To optimize edge node performance.
Answer: a
Explanation: MDP is used to model the resource allocation problem as a
sequential decision-making process

You might also like