The document proposes a novel dilution-based defense method against poisoning attacks on deep learning systems, which enhances model accuracy by adding clean data to a contaminated training dataset. Experiments demonstrate that this technique significantly reduces the success rate of these attacks and improves classification accuracy by 20.9% compared to existing defense methods. The paper outlines the challenges posed by adversarial attacks and presents a systematic approach to bolster the security of deep learning models during the training phase.
Related topics: