Purpose Image classification is perhaps the most fundamental task in imaging artificial intelligence. However, labeling images is time-consuming and tedious. We have recently demonstrated that reinforcement learning (RL) can classify 2D slices of MRI brain images with high accuracy.Here we make two important steps toward speeding image classification: Firstly, we automatically extract class labels from the clinical reports. Secondly, we extend our prior 2D classification work to fully 3D image volumes from our institution. Hence, we proceed as follows: in Part 1, we extract labels from reports automatically using a natural language processing approach termed sentence bidirectional encoder representations from transformers (SBERT). Then, in Part 2, we use these labels with RL to train a classification Deep-Q Network (DQN) for 3D image volumes.Materials and Methods For Part 1, we trained SBERT with 45 "normal" patient report impressions and 45 metastasis-containing impressions. We then used the trained SBERT to predict class labels for use in Part 2. In Part 2, we applied multi-step image classification to allow for combined Deep-Q learning using 3D convolutions and TD(0) Q learning. We trained on a set of 90 images (40 normal and 50 tumor-containing). We tested on a separate set of 61 images (40 normal and 21 tumor-containing), again using the classes predicted from patient reports by the trained SBERT in Part 1. For comparison, we also trained and tested a supervised deep learning classification network on the same set of training and testing images using the same labels.Results Part 1: Upon training with the corpus of radiology reports, the SBERT model had 100% accuracy for both normal and metastasiscontaining scans. Part 2: Then, using these labels, whereas the supervised