GloVe 100-Dimensional Word Vectors 
                Trained on
                Wikipedia and Gigaword 5 Data
              
              
             
          
        
        
          
            Released in 2014 by the computer science department at Stanford University, this representation is trained using an original method called Global Vectors (GloVe). It encodes 400,000 tokens as unique vectors, with all tokens outside the vocabulary encoded as the zero-vector. Token case is ignored.
           
        
        
        
          
            Number of layers: 1 |
          
          
            Parameter count: 40,000,100 |
          
          
            Trained size: 163 MB |
          
          
        
        
          
          Examples
          
          Resource retrieval
Get the pre-trained net:
Basic usage
Use the net to obtain a list of word vectors:
Obtain the dimensions of the vectors:
Use the embedding layer inside a NetChain:
Feature visualization
Create two lists of related words:
Visualize relationships between the words using the net as a feature extractor:
Word analogies
Get the pre-trained net:
Get a list of words:
Obtain the vectors:
Create an association whose keys are words and whose values are vectors:
Find the eight nearest words to "king":
Man is to king as woman is to:
France is to Paris as Germany is to:
Net information
Inspect the number of parameters of all arrays in the net:
Obtain the total number of parameters:
Obtain the layer type counts:
Export to MXNet
Export the net into a format that can be opened in MXNet:
Export also creates a net.params file containing parameters:
Get the size of the parameter file:
The size is similar to the byte count of the resource object:
Represent the MXNet net as a graph:
 
          
          
            Requirements
            
              
                  
                    Wolfram Language
                    
                      11.1
                    
                  
                
              
                (March 2017)
              
              
                or above
              
            
          
          
          
            Resource History
            
          
          
            Reference