Replay attacks, where recorded audio is used to spoof speaker verification systems, are a common and inexpensive threat that requires minimal technical expertise. This article presents a convolutional neural network and non-voiced audio segment-based method for detecting repeat attacks in speaker verification systems. Value on reverberation and channel noise found in non-voiced sections can help separate real audio from repeats. Emphasizing these segments' Fast Fourier Transform (FFT) spectrograms, the CNN model is trained to identify audio authenticity by minimizing the feature set by eliminating spoken components, producing faster training durations without affecting detection accuracy. Model performance is assessed using the ASVspoof 2019 dataset by the Equal Error Rate (EER) metric, obtaining an EER of 14.31% on the evaluation set and 6.75% on the development set, thereby suggesting a great capacity in separating real audio from replayed audio.