Prediction: Difference between revisions
Jump to navigation
Jump to search
No edit summary |
|||
Line 1: | Line 1: | ||
= Feature selection = | = Feature selection = | ||
* [http://r-statistics.co/Variable-Selection-and-Importance-With-R.html Feature Selection Approaches] from http://r-statistics.co | * [http://r-statistics.co/Variable-Selection-and-Importance-With-R.html Feature Selection Approaches] from http://r-statistics.co | ||
* [https://twitter.com/marktenenholtz/status/1577992133079756801 Leave-one-out feature importance (LOFO) + LightGBM] | |||
== Boruta == | == Boruta == |
Revision as of 10:29, 10 October 2022
Feature selection
- Feature Selection Approaches from http://r-statistics.co
- Leave-one-out feature importance (LOFO) + LightGBM
Boruta
Random forest
- https://en.wikipedia.org/wiki/Random_forest
- randomForest package
- Error: protect(): protection stack overflow. The trick works on my data with 26 obs and 28110 variables.
- A complete guide to Random Forest in R
- 8.5 Permutation Feature Importance from Interpretable Machine Learning by Christoph Molnar
- 15 Variable Importance from caret
- How to find the most important variables in R
- ?importance
- Random Forest Variable Importance and a simple example with 2 correlated predictors.
- If you just print the importance object from the model they are the raw importance values. However, when you use the importance function, the default for the scale argument is TRUE which returns the importance values divided by the standard error.
- Measures of variable importance in random forests
- Difference between varImp (caret) and importance (randomForest) for Random Forest
- Utilizing Machine Learning algorithms (GLMnet and Random Forest models) for Genomic Prediction of a Quantitative trait
- Gene selection and classification of microarray data using random forest 2006, and the R package varSelRF
- Feature Importance in Random Forest