DenoMAE: A Multimodal Autoencoder for Denoising Modulation Signals

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

We propose Denoising Masked Autoencoder (DenoMAE), a novel multimodal autoencoder framework for denoising modulation signals during pretraining. DenoMAE extends the concept of masked autoencoders by incorporating multiple input modalities, including noise as an explicit modality, to enhance cross-modal learning and improve denoising performance. The network is pre-trained using unlabeled noisy modulation signals and constellation diagrams, effectively learning to reconstruct their equivalent noiseless signals and diagrams. DenoMAE achieves state-of-the-art accuracy in automatic modulation classification (AMC) tasks with significantly fewer training samples, demonstrating a 10 × reduction in unlabeled pretraining data and a 3 × reduction in labeled fine-tuning data compared to existing approaches. Moreover, our model exhibits robust performance across varying Signal-to-Noise Ratios (SNRs) and supports extrapolation on unseen lower SNRs. The results indicate that DenoMAE is an efficient, flexible, and data-efficient solution for denoising and classifying modulation signals in challenging, noise-intensive environments.

Original languageEnglish
Pages (from-to)1659-1663
Number of pages5
JournalIEEE Communications Letters
Volume29
Issue number7
DOIs
StatePublished - 2025

Keywords

  • Multi-modality
  • constellation diagrams
  • denoising
  • modulation classification
  • vision transformer

Fingerprint

Dive into the research topics of 'DenoMAE: A Multimodal Autoencoder for Denoising Modulation Signals'. Together they form a unique fingerprint.

Cite this