Signal enhancement in bioacoustics can be of vital importance due to the fact that recordings are largely done in noise-heavy environments, in which anthrophonic, geophonic, or biophonic disturbances are myriad and can impede downstream analysis. Existing audio denoising techniques largely focus on human speech, whereas non-human animal vocalizations may not be compatible with frequency or temporal assumptions present in those methods. This work introduces ANIMAL–CLEAN, an animal-independent, Noise2Noise-based deep denoising toolkit utilizing a combination of signals from the target of interest as well as existing datasets used for human speech enhancement. This toolkit enables users to develop deep learning models capable of denoising bioacoustic signals with a wide range of frequencies and signal lengths, which, when used as a preprocessing step, improve results for downstream signal segmentation and classification, without the need for noise-free bioacoustic recordings.