Wolfram Function Repository
Instant-use add-on functions for the Wolfram Language
Function Repository Resource:
Fuse a BatchNormalization layer preceded by a ConvolutionLayer into a single ConvolutionLayer
ResourceFunction["NetFuseBatchNorms"][convLayer, bnLayer] fuses the weights of an initialized BatchNormalizationLayer and ConvolutionLayer into a single ConvolutionLayer whenever possible. | |
ResourceFunction["NetFuseBatchNorms"][net] repeatedly fuses the weights of an initialized BatchNormalizationLayer preceded by an initialized ConvolutionLayer into a single ConvolutionLayer in a net. |
Define a non-zero initialized ConvolutionLayer followed by a BatchNormalizationLayer:
| In[1]:= |
| Out[1]= | ![]() |
| In[2]:= | ![]() |
| Out[2]= | ![]() |
Fuse the convLayer and bnLayer into a single ConvolutionLayer:
| In[3]:= |
| Out[3]= | ![]() |
Create a NetChain:
| In[4]:= |
| Out[4]= | ![]() |
Perform the same combination within a NetChain:
| In[5]:= |
| Out[5]= | ![]() |
Compare the outputs on a random input:
| In[6]:= |
| Out[7]= |
Note that the fused layer is faster:
| In[8]:= |
| Out[8]= |
| In[9]:= |
| Out[9]= |
Note that NetFuseBatchNorms does not do anything on an uninitialized net:
| In[10]:= | ![]() |
| Out[10]= | ![]() |
| In[11]:= |
| Out[11]= | ![]() |
| In[12]:= |
| Out[12]= |
Get pretrained ShuffleNet-V2:
| In[13]:= | ![]() |
Accelerate ShuffleNet-V2:
| In[14]:= |
Compare the outputs on a random image:
| In[15]:= |
| Out[16]= |
The fused net is lighter as it contains less BatchNormalizationLayer layers:
| In[17]:= |
| Out[17]= |
| In[18]:= |
| Out[18]= |
This work is licensed under a Creative Commons Attribution 4.0 International License