Home

Awesome

Double-Attention-Network

This is the PyTorch implementation of A^2-Nets: Double Attention Networks, Y Chen et al NIPS 2018

It can be used as an additional block for building models. Right now, the output tensor has shape (B, c_n, H, W). One can re-construct the original shape (B, c, H, W) with a single line of code in PyTorch.

Layer architecture