A) In this exercise, we selected the RMSprop optimizer. The RMSprop is a good choice because it allows for an increased learning rate which allows the alogirthm to “take larget steps”, and thus, converge faster! In RMSprop, the learning rate is adapted by dividing by the root of the sequared gradient. I think RMSprop is better for small batches in large datasets compared to other optimizers like adam.
B) The selected loss function is binary crossentropy. The loss function will return high values for bad predictions of binary classification, and low values for good predictions.
C) In the code, when compiling the model, it says: metric = [accuracy]. Assigning this class creates 2 variables: total and count. Total is the number of times y_pred is equal to y_true in a binary classiication problem. In the end, total is divided by count to calculate frequency.
As epochs increased, validation loss increased while training loss increased - showing that the model may be overfit. With respect to accuracy, they both increased as epochs icnreased, and training accuracy was slightly higher which reveals the model is slightly overfit. Overall, I would say the model performed well.
Below, are the dog images I tested in the CNN.