From 515f1d473c6c80279beb8d2e737ea73f178ac855 Mon Sep 17 00:00:00 2001 From: Bruhat Musunuru <61818239+BruhatM@users.noreply.github.com> Date: Sat, 12 Dec 2020 02:03:02 -0800 Subject: [PATCH] Update reports.md --- reports/reports.md | 10 +++------- 1 file changed, 3 insertions(+), 7 deletions(-) diff --git a/reports/reports.md b/reports/reports.md index b0e2f4d..8dcb991 100644 --- a/reports/reports.md +++ b/reports/reports.md @@ -184,11 +184,7 @@ from physicochemical test Since this is a multi-class classification, our goal was to find a model -that was consistent and able to recognize patterns from our data. We -choose to use a neutral network Multi-layer Perception (MLP) model as it -was consistent and showed promising results. If we take a look at the -accuracy scores and f1 scores across cross validation splits, we can see -that it is pretty consistent which was not the case with many models. +that was consistent and able to recognize patterns from our data. After looking at the results, MLP and random forests seem to be the best models for our data. But after analysing the models, We concluded that Random forests is picking up noise in the data as the train error is comparatively drastically lower than the validation error. This is a sign of over fitting which does not seem to affect MLP. Furthermore, the cross-validation scores are more consistent and normally distributed compared to Random forests which further cements that MLP is the best pick.