As artificial intelligence models continue to grow in size, the demand for computing and storage resources has increased significantly. Optimizing artificial intelligence models on GPUs to fully utilize the available resources is currently a major challenge in the field of High-Performance Computing (HPC). Graph structures in computers are commonly stored in matrix form, and to conserve computer storage resources, Compressed Sparse Row (CSR) or Compressed Sparse Column (CSC) matrix compression is often employed. Previous GPGPU Prefetcher studies have primarily focused on thread-aware prefetching. However, due to the complexity of graph structures, it is challenging to determine the storage location of the next traversed node during traversal. Therefore, thread-aware prefetching methods are not suitable for graph structures. In this work, we address the application of graph structures in GPGPU architecture and design two new Prefetchers (C-Prefetcher and X-Prefetcher) based on CSR and CSC. When the SIMT Core loads Cache Line from Global memory to L2 Cache, C-Prefetcher fetches the next accessed data to L2 Cache based on CSC or CSR. Building upon C-Prefetcher, This work further propose X-Prefetcher. Through experimental comparisons in this paper involving multiple graph applications, the results indicate that the LLC (Last Level Cache) Miss rates during graph traversal are respectively reduced by 4.5% and 11% compared to the Baseline architecture, where no Prefetcher is employed, for C-Prefetcher and X-Prefetcher architectures. Additionally, the IPC (Instructions Per Cycle) is increased by 3.2% and 6.8% for C-Prefetcher and X-Prefetcher architectures, respectively, compared to the Baseline architecture during graph traversal.