Fig. 1 | Scientific Reports

Fig. 1

From: Efficient attention vision transformers for monocular depth estimation on resource-limited hardware

Fig. 1

The general framework used to compare the different network-optimization configurations. The module Network architecture instantiates the network to be analysed, while the submodule Efficient Attention Module integrates the attention mechanism used by the encoder, decoder, or the entire network. On the left side a comparison is shown to illustrate how each optimization modifies the original module. The left module depicts the optimised attention, while the right one shows its standard version.

Back to article page