@@ -304,6 +304,8 @@ We have conducted a series of experiments to evaluate the performance of differe
🔍 For our basic classifiers (Naive Bayes, Decision Tree, Random Forest), we have utilized scikit-learn's `GridSearchCV` to optimize the hyperparameters. This approach allows us to systematically test different hyperparameter combinations and identify the best configuration for each classifier. In order to use the development set for hyperparameter tuning, we have used the `PredefinedSplit` function to create a fixed split between the training and development set instead of using a default cross-validation split.
For the random forest classifier, we have conducted additional experiments involving various picture sizes and number of estimators. This exploration was motivated by the notably promising results obtained from this classifier compared to the other basic classifiers.
↪️ To reproduce the results, please refer to the [README](src/README.md) in the `src` folder for instructions on how to run the experiments.
### Feature Engineering
...
...
@@ -422,8 +424,8 @@ The following table shows an excerpt of the feature and size combination used:
| Resized | Features | Accuracy (Dev) | Best Parameters | Comments |