Resource allocation is a critical task in 5Gnetworks that determines how network resources are assigned to different devices and services. Traditional methods rely on predefined rules or heuristics, which may not always be optimal. Deep reinforcement learning (DRL)is a promising approach for radio resource allocation in 5Gnetworks as it can learn to optimize resource allocation based on feedback from the network. In DRL, an agent learns to make decisions based on rewards and penalties received from the environment. In radio resource allocation, the agent would learn to allocate resources, such as frequency bands and power levels, to different devices and services to maximize some performance metric, such asthroughput or energy efficiency. The main challenge in applying DRL to radio resource allocation is designing an appropriate reward function that incentivizes the agent to improve the performance metric while avoiding undesirable behavior. Additionally, the radio resource allocation problem is complex, requiring the agent to consider many variables and constraints, such as channel conditions, interference, and QoS requirements. To address this, researchers have proposed various techniques such as hierarchical RL, multi-agent RL, and curriculum learning. Despite the challenges, DRL has shown promising results inradio resource allocation for 5G networks. It has outperformed traditional methods in some scenarios, especially when network conditions are dynamic and unpredictable. However, further research is necessary to explore the scalability and robustness of DRL-based approaches in practical 5G networks. In this method we suggest an algorithm for voice and data carriers in sub-6 GHz and millimeter wave (mmWave) frequencies respectively. The mmWave ranges between 30GHz to 300GHz