GIANT Networks: Very Deep Fully Connected Neural Networks Applied to Microwave Problems
Journal article, 2026

We present the Gradient-Informed Attentive Normalisation Training (GIANT) framework with the objective to create very deep fully connected neural networks, which we use as surrogate models in the context of microwave problems. As the central component of the GIANT framework, we introduce a novel dynamic reparameterisation procedure for the weight-bias parameter space by means of a low-variance preserving normalisation layer for each fully connected layer and we refer to this construction as Attentive Normalisation (AttNorm). As part of AttNorm, we also introduce a new and tailored updating scheme that improves the convergence during training. To efficiently train very deep fully connected neural networks, we exploit Sobolev training with gradient information, which is computed at a very low computational cost by means of continuum sensitivity analysis. We test our novel approach on two microwave applications: (i) a six-port microwave cavity with a random medium and (ii) an H-plane waveguide filter optimised under geometrical uncertainty. For these examples, we demonstrate successful training of neural networks with up to 30 layers, which are sufficiently accurate and expressive to serve as excellent surrogate models.

microwave resonators

microwave filters

optimisation

sensitivity analysis

neural nets

Author

Simon Stenmark

Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering

Thomas Rylander

Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering

Tomas McKelvey

Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering

Andrei Osipov

Chalmers, Space, Earth and Environment, Astronomy and Plasmaphysics

IET Microwaves, Antennas and Propagation

1751-8725 (ISSN) 17518733 (eISSN)

Vol. 20 1 e70077

Areas of Advance

Information and Communication Technology

Subject Categories (SSIF 2025)

Communication Systems

Computer Vision and learning System

Computer Sciences

DOI

10.1049/mia2.70077

More information

Latest update

3/16/2026