Keras AttentiveNormalization Save Abandoned

Unofficial Keras implementation of the paper Attentive Normalization.

Project README

Attentive Normalization

This repository is an unofficial Keras implementation of the paper Attentive Normalization by Xilai Li, Wei Sun and Tianfu Wu.

The official implementation will be released here : https://github.com/ivMCL/

Introduction

Attentive Normalization (AN) is an attention-based version of BN which recalibrates channel information of BN. AN absorbs the Squeeze-and-Excitation (SE) mechanism into the affine transformation of BN. AN learns a small number of scale and offset parameters per channel (i.e., different affine transformations). Their weighted sums (i.e., mixture) are used in the final affine transformation. The weights are instance-specific and learned in a way that channel-wise attention is considered, similar in spirit to the squeeze module in the SE unit. This can be used as a drop-in replacement of standard BatchNormalization layer.

Usage

Please refer to the notebook for an usage example.

Open Source Agenda is not affiliated with "Keras AttentiveNormalization" Project. README Source: Cyril9227/Keras_AttentiveNormalization
Stars
29
Open Issues
0
Last Commit
4 years ago
License
MIT

Open Source Agenda Badge

Open Source Agenda Rating