Image-based radio frequency fingerprinting (RFF) is a promising variant of traditional RFF systems. As a distinctive feature, such systems convert physical-layer signals into matrices resembling 2-D or 3-D images and consider the latter as the input for state-of-the-art image classifiers. Compared to traditional ones, image-based RFF systems have recently shown enhanced flexibility for device identification as they can better mitigate channel conditions, devices movement, and power cycle. However, previous works have yet to investigate their performance when subject to adversarial machine learning (AML) attacks using state-of-the-art techniques such as generative adversarial networks and the fast gradient sign method. Similarly, there are no studies on their capability to integrate adversarial learning strategies for enhancing their robustness to such attacks. In this article, we fill the gap by conducting an experimental analysis of the effectiveness of AML attacks and adversarial training techniques for image-based RFF systems. Using a state-of-the-art image-based RFF system and actual measurements, we show that adversarial samples can effectively degrade classification performance. At the same time, training the image-based RFF system with adversarial samples increases the reliability and robustness of such methods at the cost of a lower classification accuracy.
Adversarial Machine Learning for Image-Based Radio Frequency Fingerprinting: Attacks and Defenses / Papangelo, Lorenzo; Pistilli, Maurizio; Sciancalepore, Savio; Oligeri, Gabriele; Piro, Giuseppe; Boggia, Gennaro. - In: IEEE COMMUNICATIONS MAGAZINE. - ISSN 0163-6804. - STAMPA. - 62:11(2024), pp. 108-113. [10.1109/MCOM.001.2300464]
Adversarial Machine Learning for Image-Based Radio Frequency Fingerprinting: Attacks and Defenses
Giuseppe Piro;Gennaro Boggia
2024-01-01
Abstract
Image-based radio frequency fingerprinting (RFF) is a promising variant of traditional RFF systems. As a distinctive feature, such systems convert physical-layer signals into matrices resembling 2-D or 3-D images and consider the latter as the input for state-of-the-art image classifiers. Compared to traditional ones, image-based RFF systems have recently shown enhanced flexibility for device identification as they can better mitigate channel conditions, devices movement, and power cycle. However, previous works have yet to investigate their performance when subject to adversarial machine learning (AML) attacks using state-of-the-art techniques such as generative adversarial networks and the fast gradient sign method. Similarly, there are no studies on their capability to integrate adversarial learning strategies for enhancing their robustness to such attacks. In this article, we fill the gap by conducting an experimental analysis of the effectiveness of AML attacks and adversarial training techniques for image-based RFF systems. Using a state-of-the-art image-based RFF system and actual measurements, we show that adversarial samples can effectively degrade classification performance. At the same time, training the image-based RFF system with adversarial samples increases the reliability and robustness of such methods at the cost of a lower classification accuracy.File | Dimensione | Formato | |
---|---|---|---|
2024_Adversarial_Machine_Learning_for_Image-Based_Radio_Frequency_Fingerprinting_postprint.pdf
accesso aperto
Tipologia:
Documento in Post-print
Licenza:
Tutti i diritti riservati
Dimensione
978.92 kB
Formato
Adobe PDF
|
978.92 kB | Adobe PDF | Visualizza/Apri |
2024_Adversarial_Machine_Learning_for_Image-Based_Radio_Frequency_Fingerprinting_Attacks_and_Defenses.pdf
solo gestori catalogo
Tipologia:
Versione editoriale
Licenza:
Tutti i diritti riservati
Dimensione
705.69 kB
Formato
Adobe PDF
|
705.69 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.