Wolfram Neural Net Repository
Immediate Computable Access to Neural Net Models
Identify the main object in an image
Released in 2018, this model features pointwise group convolutions and bottleneck-like structures. A "channel shuffle" operation is introduced to enable information flow between different groups of channels and improve accuracy.
Number of layers: 243 | Parameter count: 2,294,784 | Trained size: 10 MB |
This model achieves a top-1 accuracy of 69.4% on the original ImageNet validation set.
Get the pre-trained net:
In[1]:= | ![]() |
Out[1]= | ![]() |
Classify an image:
In[2]:= | ![]() |
Out[2]= | ![]() |
The prediction is an Entity object, which can be queried:
In[3]:= | ![]() |
Out[3]= | ![]() |
Get a list of available properties of the predicted Entity:
In[4]:= | ![]() |
Out[4]= | ![]() |
Obtain the probabilities of the 10 most likely entities predicted by the net:
In[5]:= | ![]() |
Out[5]= | ![]() |
An object outside the list of the ImageNet classes will be misidentified:
In[6]:= | ![]() |
Out[6]= | ![]() |
Obtain the list of names of all available classes:
In[7]:= | ![]() |
Out[7]= | ![]() |
Remove the last layers of the trained net so that the net produces a vector representation of an image:
In[8]:= | ![]() |
Out[8]= | ![]() |
Get a set of images:
In[9]:= | ![]() |
Visualize the features of a set of images:
In[10]:= | ![]() |
Out[10]= | ![]() |
Extract the weights of the first convolutional layer in the trained net:
In[11]:= | ![]() |
Show the dimensions of the weights:
In[12]:= | ![]() |
Out[12]= | ![]() |
Visualize the weights as a list of 24 images of size 3⨯3:
In[13]:= | ![]() |
Out[13]= | ![]() |
Use the pre-trained model to build a classifier for telling apart images of sunflowers and roses. Create a test set and a training set:
In[14]:= | ![]() |
In[15]:= | ![]() |
Remove the last layers from the pre-trained net:
In[16]:= | ![]() |
Out[16]= | ![]() |
Create a new net composed of the pre-trained net followed by a linear layer and a softmax layer:
In[17]:= | ![]() |
Out[17]= | ![]() |
Train on the dataset, freezing all the weights except for those in the "linearNew" layer (use TargetDevice -> "GPU" for training on a GPU):
In[18]:= | ![]() |
Out[18]= | ![]() |
Accuracy obtained on the test set:
In[19]:= | ![]() |
Out[19]= | ![]() |
Inspect the number of parameters of all arrays in the net:
In[20]:= | ![]() |
Out[20]= | ![]() |
Obtain the total number of parameters:
In[21]:= | ![]() |
Out[21]= | ![]() |
Obtain the layer type counts:
In[22]:= | ![]() |
Out[22]= | ![]() |
Display the summary graphic:
In[23]:= | ![]() |
Out[23]= | ![]() |
Export the net to the ONNX format:
In[24]:= | ![]() |
Out[24]= | ![]() |
Get the size of the ONNX file:
In[25]:= | ![]() |
Out[25]= | ![]() |
The size is similar to the byte count of the resource object:
In[26]:= | ![]() |
Out[26]= | ![]() |
Check some metadata of the ONNX model:
In[27]:= | ![]() |
Out[27]= | ![]() |
Import the model back into the Wolfram Language. However, the NetEncoder and NetDecoder will be absent because they are not supported by ONNX:
In[28]:= | ![]() |
Out[28]= | ![]() |
Wolfram Language 12.3 (May 2021) or above