Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Actually, no [1]. From the abstract:

We can cause the network to misclassify an image by applying a certain imperceptible perturbation, which is found by maximizing the network's prediction error. In addition, the specific nature of these perturbations is not a random artifact of learning: the same perturbation can cause a different network, that was trained on a different subset of the dataset, to misclassify the same input.

[1] https://arxiv.org/abs/1312.6199

EDIT: Also interesting: "Universal Adversarial Perturbations" https://arxiv.org/abs/1610.08401




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: