···149149 (binary) tree. Each non-leaf node contain a selection criteria to its
150150 branches. Every leaf node contains the class that will be assigned to the
151151 instance if the node is reached. In other training methods, decision trees
152152- have the tendency to overfit, but in random forest a multitude of decision
153153- tree is trained with a certain degree of randomness and the mean of these
154154- trees is used which avoids this problem.
152152+ have the tendency to overfit\footnote{Overfitting occurs when a model learns
153153+ from the data too specifically, and loses its ability to generalise its
154154+ predictions for new data (resulting in loss of prediction accuracy)}, but in
155155+ random forest a multitude of decision tree is trained with a certain degree
156156+ of randomness and the mean of these trees is used which avoids this problem.
155157156158 \subsection{Neural Network Architectures}
157159···233235 models; requiring training on a dataset of typical images. Each network was
234236 trained using the preprocessed training dataset and labels, for 25 epochs
235237 (one forward and backward pass of all data) in batches of 150. The number of
236236- epochs was chosen to maximise training time and prevent
237237- overfitting\footnote{Overfitting occurs when a model learns from the data
238238- too specifically, and loses its ability to generalise its predictions for
239239- new data (resulting in loss of prediction accuracy)} of the training data,
240240- given current model parameters. The batch size is the number of images sent
241241- through each pass of the network. Using the entire dataset would train the
242242- network quickly, but decrease the network's ability to learn unique features
243243- from the data. Passing one image at a time may allow the model to learn more
244244- about each image, however it would also increase the training time and risk
245245- of overfitting the data. Therefore the batch size was chosen to maintain
246246- training accuracy while minimising training time.
238238+ epochs was chosen to maximise training time and prevent overfitting of the
239239+ training data, given current model parameters. The batch size is the number
240240+ of images sent through each pass of the network. Using the entire dataset
241241+ would train the network quickly, but decrease the network's ability to learn
242242+ unique features from the data. Passing one image at a time may allow the
243243+ model to learn more about each image, however it would also increase the
244244+ training time and risk of overfitting the data. Therefore the batch size was
245245+ chosen to maintain training accuracy while minimising training time.
247246248247 \subsection{Neural Network Testing}\label{nnTesting}
249248