Evaluation of Momentum Diverse Input Iterative Fast Gradient Sign Method (M-DI2-FGSM) Based Attack Method on MCS 2018 Adversarial Attacks on Black Box Face Recognition System

June 23, 2018 ยท Entered Twilight ยท ๐Ÿ› arXiv.org

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE, README.md, Welcome_notebook.ipynb, __init__.py, attacker.py, attacker_1.py, attacker_2.py, attacker_3.py, attacker_4.py, attacker_foolbox_1.py, attacker_model.py, attacks, class_name_attacks.py, code, data_augmentor.ipynb, downloader.py, evaluate.py, foolbox_attack_pytorch.ipynb, foolbox_attack_pytorch.py, gpu_model.py, make_attack.sh, make_data_train.sh, make_data_val.sh, make_evaluate_0.sh, make_evaluate_1.sh, make_train.sh, prepare_data.py, requirements.txt, student_net_learning, submit.ipynb, utils, welcome.py

Authors Md Ashraful Alam Milton arXiv ID 1806.08970 Category cs.CV: Computer Vision Citations 16 Venue arXiv.org Repository https://github.com/miltonbd/mcs_2018_adversarial_attack โญ 16 Last Checked 2 months ago
Abstract
The convolutional neural network is the crucial tool for the recent success of deep learning based methods on various computer vision tasks like classification, segmentation, and detection. Convolutional neural networks achieved state-of-the-art performance in these tasks and every day pushing the limit of computer vision and AI. However, adversarial attack on computer vision systems is threatening their application in the real life and in safety-critical applications. Necessarily, Finding adversarial examples are important to detect susceptible models to attack and take safeguard measures to overcome the adversarial attacks. In this regard, MCS 2018 Adversarial Attacks on Black Box Face Recognition challenge aims to facilitate the research of finding new adversarial attack techniques and their effectiveness in generating adversarial examples. In this challenge, the attack"s nature is targeted-attack on the black-box neural network where we have no knowledge about black-block"s inner structure. The attacker must modify a set of five images of a single person so that the neural network miss-classify them as target image which is a set of five images of another person. In this competition, we applied Momentum Diverse Input Iterative Fast Gradient Sign Method (M-DI2-FGSM) to make an adversarial attack on black-box face recognition system. We tested our method on MCS 2018 Adversarial Attacks on Black Box Face Recognition challenge and found competitive result. Our solution got validation score 1.404 which better than baseline score 1.407 and stood 14 place among 132 teams in the leader-board. Further improvement can be achieved by finding improved feature extraction from source image, carefully chosen hyper-parameters, finding improved substitute model of the black-box and better optimization method.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision