https://www.mdu.se/

mdu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Learning Activation Functions for Adversarial Attack Resilience in CNNs
Mälardalen University, School of Innovation, Design and Engineering, Embedded Systems.
Mälardalen University, School of Innovation, Design and Engineering, Embedded Systems.
Mälardalen University, School of Innovation, Design and Engineering, Embedded Systems.
2023 (English)In: Lect. Notes Comput. Sci., Springer Science and Business Media Deutschland GmbH , 2023, p. 203-214Conference paper, Published paper (Refereed)
Abstract [en]

Adversarial attacks on convolutional neural networks (CNNs) have been a serious concern in recent years, as they can cause CNNs to produce inaccurate predictions. Through our analysis of training CNNs with adversarial examples, we discovered that this was primarily caused by naïvely selecting ReLU as the default choice for activation functions. In contrast to the focus of recent works on proposing adversarial training methods, we study the feasibility of an innovative alternative: learning novel activation functions to make CNNs more resilient to adversarial attacks. In this paper, we propose a search framework that combines simulated annealing and late acceptance hill-climbing to find activation functions that are more robust against adversarial attacks in CNN architectures. The proposed search method has superior search convergence compared to commonly used baselines. The proposed method improves the resilience to adversarial attacks by achieving up to 17.1%, 22.8%, and 16.6% higher accuracy against BIM, FGSM, and PGD attacks, respectively, over ResNet-18 trained on the CIFAR-10 dataset.

Place, publisher, year, edition, pages
Springer Science and Business Media Deutschland GmbH , 2023. p. 203-214
Series
Lecture Notes in Computer Science, ISSN 0302-9743 ; 14125 LNAI
Keywords [en]
Activation Function, Adversarial Attack, Convolutional Neural Network, Robustness, Activation analysis, Chemical activation, Convolution, Convolutional neural networks, Activation functions, Attack resiliences, High-accuracy, Hill climbing, Neural network architecture, Search method, Training methods, Simulated annealing
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:mdh:diva-64441DOI: 10.1007/978-3-031-42505-9_18Scopus ID: 2-s2.0-85172420687ISBN: 9783031425042 (print)OAI: oai:DiVA.org:mdh-64441DiVA, id: diva2:1802768
Conference
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Available from: 2023-10-05 Created: 2023-10-05 Last updated: 2025-11-21Bibliographically approved
In thesis
1. Enhancing Perception System Robustness Against Attacks and Natural Perturbations
Open this publication in new window or tab >>Enhancing Perception System Robustness Against Attacks and Natural Perturbations
2026 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Deep learning has led to major progress in computer vision, but modern Deep Neural Networks (DNNs) are still highly vulnerable to input perturbations, which limits their robustness in safety-critical applications. This challenge becomes even more critical in real-world industrial environments, such as autonomous machinery operating on construction sites, where visual data is influenced by unpredictable weather conditions, variable lighting, and physical wear and degradation. In addition, data scarcity, privacy constraints, and domain shift prevent the direct application of conventional large-scale training pipelines. 

 

This thesis addresses these challenges by proposing a comprehensive, multi-level framework that strengthens model-level robustness against adversarial attacks, enhances data-level robustness to natural environmental perturbations, and improves adaptive learning under distributed and data-constrained conditions, enabling reliable deployment of visual perception models in complex, safety-critical environments.

 

The first contribution focuses on the robustness of model-level attacks against adversarial attacks. A meta-heuristic search method is proposed to automatically discover activation functions that increase resistance to adversarial perturbations without requiring adversarial training. A hybrid search strategy further improves convergence efficiency, yielding Convolutional Neural Networks (CNNs) that outperform standard architectures under adversarial attacks while maintaining competitive clean-data accuracy.

 

The second contribution introduces ConstScene, a large-scale semantic segmentation dataset representing real and synthetic construction-site imagery under diverse weather and sensor degradation conditions. Experiments reveal significant performance drops when models trained on clean data are exposed to perturbed inputs, demonstrating the need for environment-specific robustness benchmarks.

 

The third contribution introduces an integrated framework that combines Federated Learning (FL) for decentralized collaborative training with Few-Shot Learning (FSL) for sample-efficient domain adaptation, supported by server-side Hyperparameter Optimization (HPO). The proposed approach enables effective model adaptation across distributed construction sites without sharing raw data, significantly improving robustness across heterogeneous client datasets.

 

In general, this thesis proposes three contributions to enhance robustness in perception systems: model-level robustness against adversarial attacks, introducing the ConstScene dataset for benchmarking performance under real-world degradations and data-level robustness against natural perturbations, and an integrated framework enabling decentralized, sample-efficient model adaptation across heterogeneous environments.

Place, publisher, year, edition, pages
Västerås: Mälardalen University, 2026. p. 225
Series
Mälardalen University Press Dissertations, ISSN 1651-4238 ; 452
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:mdh:diva-74467 (URN)978-91-7485-736-8 (ISBN)
Public defence
2026-01-23, Gamma, Mälardalens universitet, Västerås, 10:00 (English)
Opponent
Supervisors
Available from: 2025-11-25 Created: 2025-11-21 Last updated: 2026-01-02Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Salimi, MaghsoodLoni, MohammadSirjani, Marjan

Search in DiVA

By author/editor
Salimi, MaghsoodLoni, MohammadSirjani, Marjan
By organisation
Embedded Systems
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 142 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf