Abstract:
Deep neural networks play a vital role in image recognition. There are so many
mission-critical applications that use deep neural networks for image recognition. With
the popularization of deep neural networks, attackers have identified their downsides
of them when it comes to image recognition. Some ways can create images that can
fool even deep neural networks. These images are commonly known as adversarial
images. So attackers use these adversarial images to fool image recognition neural
networks to develop a negative picture about using neural networks for image
recognition. And even sometimes, attackers use these loopholes to conduct criminal
activities as well. Keeping all these aspects in mind the idea of the research is to
develop a viable solution that can tackle the main two attack techniques. The research
will focus on developing adversarial images using main attacking techniques and
developing a defense mechanism for those attacks. The defense technique used in the
research is a combination of two techniques called adversarial training and defense
distillation. As the outcome of the project accuracy of the proposed solution is
measured against a typical deep neural network-based image recognition system using
data samples containing adversarial images.
Citation:
Amarasinghe, P.T. (2022). Handling adversaries in image recognition deep neural networks [Master's theses, University of Moratuwa]. Institutional Repository University of Moratuwa.http://dl.lib.uom.lk/handle/123/22410