Page 45 - Chip Scale Review_May June_2023-digital
P. 45

Figure 2: CNN structure of the proposed modified AlexNet neural network. SOURCE: Created using the NN-SVG diagramming tool, (www.github.com/alexlenail/NN-SVG).
                                                                              value or the average value for a group
                                                                              of adjacent pixels, which causes loss of
                                                                              information, so it is abandoned because
                                                                              it could reduce the test accuracy of the
                                                                              shmoo result types Hole and Marginal.
                                                                                Additionally, the batch normalization
                                                                              layer is implemented instead of using
                                                                              t he  d ropout  met hod  to  suppress
                                                                              overfitting. With batch normalization
                                                                              the  data  is  standardized  in  a  mini-
                                                                              batch,  with  the  mean  value  of  the
                                                                              nor malized  data  being  0  and  the
                                                                              standard deviation as 1. This process
                                                                              is similar to dropout, as it “discards”
                                                                              a part of the nodes where the output is
                                                                              close to 0 at this layer. It also makes
                                                                              the output of each layer follow the
                                                                              same distribution, and so, as a result,
        Figure 3: Diagram of a spatial pyramid pooling layer.
                                                                              eliminates the potential of “parameter
          A significant difference between   point is processed as a pixel of the   explosion” and “parameter attenuation”
        AlexNet and this neural network is   input shmoo image, so the result of the   in the deep network structure training
        that all pooling layers except the SPP   shmoo is sensitive to every test point.   process. This stabilizes and accelerates
        are removed. This is because each test   The pooling layer takes the maximum   the model’s training [5].



























        Figure 4: Shmoo samples in the training dataset.

                                                                                                             43
                                                               Chip Scale Review   May  •  June  •  2023   [ChipScaleReview.com]  43
   40   41   42   43   44   45   46   47   48   49   50