Capitule 6. Classifiers

Choose a classifier sometimes is a difficult task. It depends on the problem to resolve, and also it is essential to know the "no free lunch" theorem that said: "On the criterion of generalization performance, there are no context -or problem- independent reasons to favor one learning or classification method to another."
Most of the classifiers need parameters, the "correct" configuration of these parameters can improve the final performance of our model. Although in practice, we need to consider other techniques such as feature engineering to improve performance.
Standard classifiers in the literature are mentioned in the next list:

  • Support Vector Machine (SVM)
  • Random Forest
  • XGBoost
  • Artificial Neural Network (ANN)
  • K-Nearest Neighbour (KNN)
  • Gaussian Naive Bayes
The classifier to use will depend on the type of dataset and the area. In general, the performance of the Random Forest and XGBoost is not bad.

THE TASK FOR YOU:
  • Which are the parameters for the Random Forest and XGBoost classifiers?
  • Which is the most used language programming for machine learning?
  • Write the syntaxis to invoke the classifiers of the first point (use the Scikit-learn tool).