This paper discusses an approach to target detection and classification based on acoustic signals collected using one single microphone. This study has applications to sonar or any other sound event classification system. We divide the problem into two parts, namely feature extraction and target detection and classification. We use an optimization step based on human auditory uncertainty. We employ a majority voting rule for every set of feature vectors, i.e., an estimate is only performed if the majority agrees. We conducted experiments using a single channel of the AIRA-UAS dataset, a public database of raw drone noises collected with an array of microphones mounted on a drone. This dataset comprises many different kinematics, with different spectra. The features we used are based on the Mel-Frequency Cepstral Coefficients (MFCC) and the Short-Time Fourier Transform of raw signals. We used the K-Nearest Neighbors algorithm for classification and adopted the crossvalidation strategy to evaluate the method. We observed that the use of MFCC results in less biased estimations, which favors the voting strategy. The detection in the proposed method reached a probability of false positive near 0%, even with a small set of votes, and a classification accuracy of 99.1%. These metrics satisfy the requirements of most civilian and military applications.