Feature Preparation, Selection and Engineering
In the previous lesson, we made our first submission to Kaggle. Kaggle is a site where people create algorithms and compete against machine learning practitioners around the world. Your algorithm wins the competition if it's the most accurate on a particular data set. Using Kaggle and this Kaggle Fundamentals course, you will have a fun way to practice your machine learning skills.
In this lesson, we're going to focus on working with the features used in the model to boost the accuracy of our predictions. To do this, we'll start by looking at feature selection. Feature selection is important because it helps to exclude features which are not good predictors, or features that are closely related to each other.
As you work through each concept, you’ll get to apply what you’ve learned from within your browser so that there's no need to use your own machine to do the exercises. The Python environment inside of this course includes answer checking so you can ensure that you've fully mastered each concept before learning the next concept.
2. Preparing More Features
3. Determining the Most Relevant Features
4. Training a model using relevant features.
5. Submitting our Improved Model to Kaggle
6. Engineering a New Feature Using Binning
7. Engineering Features From Text Columns
8. Finding Correlated Features
9. Final Feature Selection using RFECV
10. Training A Model Using our Optimized Columns
11. Submitting our Model to Kaggle
12. Next Steps