calculates the dot products between the input and the weights and adds the bias, it
103
The general formula for softmax is shown in Figure
3-43
.
As for the
optimizer,
it is set to the Adam optimizer,
a type of gradient-based
optimizer. By default, the parameter known as the
learning rate is set to 0.001. Recall
that the learning rate helps determine the step size taken by the optimization algorithm
to see how much to adjust the weights by.
After executing the code in Figure
3-43
, you get the output in Figure
3-44
.
Do'stlaringiz bilan baham: