This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
en:iot-reloaded:random_forests [2024/12/10 16:12] – blanka | en:iot-reloaded:random_forests [2024/12/10 21:39] (current) – pczekalski | ||
---|---|---|---|
Line 1: | Line 1: | ||
- | ====== Random | + | ====== Random |
- | Random forests ((https:// | + | Random forests ((https:// |
- | < | + | < |
{{ : | {{ : | ||
< | < | ||
</ | </ | ||
- | Now, one might consider several forecast agents – friends of neighbours - where each provides their own forecast depending on the factor values. Some will be higher than the actual value, and some lower. However, since they all use some **experience-based knowledge**, | + | Now, one might consider several forecast agents – friends of neighbours - where each provides their forecast depending on the factor values. Some will be higher than the actual value, and some will be lower. However, since they all use some **experience-based knowledge**, |
- | The Random forest (RF) method uses hundreds of forecast agents, and decision trees, and then applies majority voting. | + | The Random forest (RF) method uses hundreds of forecast agents and decision trees and then applies majority voting |
- | < | + | < |
{{ : | {{ : | ||
< | < | ||
Line 18: | Line 18: | ||
Some advantages: | Some advantages: | ||
* RF uses more knowledge than a single decision tree. | * RF uses more knowledge than a single decision tree. | ||
- | * Furthermore, | + | * Furthermore, |
* This is true because a single data source might suffer from data anomalies reflected in model anomalies. | * This is true because a single data source might suffer from data anomalies reflected in model anomalies. | ||
Line 31: | Line 31: | ||
* If the number of cases in the training set is N, a sample of N cases at random is taken - but with replacement, | * If the number of cases in the training set is N, a sample of N cases at random is taken - but with replacement, | ||
* This sample will be the training set for growing the tree. | * This sample will be the training set for growing the tree. | ||
- | * If there are M input factors, a number m<<M (m is significantly smaller than M) is specified such that at each node, m factors are selected | + | * If there are M input factors, a number m<<M (m is significantly smaller than M) is specified such that at each node, m factors are selected |
* The value of m is held constant while the forest grows. | * The value of m is held constant while the forest grows. | ||
* Each tree is grown to the largest extent possible. **There is no pruning**. | * Each tree is grown to the largest extent possible. **There is no pruning**. | ||
Line 38: | Line 38: | ||
===== Additional considerations ===== | ===== Additional considerations ===== | ||
- | **Correlation Between Trees in the Forest:** The correlation between any two trees in a Random Forest refers to the similarity in their predictions across the same dataset. When trees are highly correlated, they are likely | + | **Correlation Between Trees in the Forest:** The correlation between any two trees in a Random Forest refers to the similarity in their predictions across the same dataset. When trees are highly correlated, they will likely make similar mistakes on the same inputs. In other words, if many trees make similar errors, the model' |
* **Bootstrap Sampling:** Each tree is trained on a different bootstrapped sample (random sampling with replacement) of the training data, which helps to reduce the correlation between the trees. | * **Bootstrap Sampling:** Each tree is trained on a different bootstrapped sample (random sampling with replacement) of the training data, which helps to reduce the correlation between the trees. | ||
* **Feature Randomness: | * **Feature Randomness: | ||
Line 45: | Line 45: | ||
**Strength of Each Individual Tree:** The strength of an individual tree refers to its classification accuracy on new data, i.e., its ability to perform as a string classifier. In Random Forest terminology, | **Strength of Each Individual Tree:** The strength of an individual tree refers to its classification accuracy on new data, i.e., its ability to perform as a string classifier. In Random Forest terminology, | ||
- | Each tree's strength depends on various factors, including its depth and the features it uses for splitting. However, there is a trade-off between correlation and strength. For example, reducing m (the number of features considered at each split) increases the diversity among the trees, lowering correlation, | + | Each tree's strength depends on various factors, including its depth and the features it uses for splitting. However, there is a trade-off between correlation and strength. For example, reducing m (the number of features considered at each split) increases the diversity among the trees, lowering correlation. Still, it may also reduce the strength of each tree, as it may limit its access to highly predictive features. |
- | Despite this trade-off, Random Forests balance these dynamics by optimising m to minimise the ensemble error. Generally, a moderate reduction in m lowers correlation without significantly compromising the strength of each tree, thus leading to an overall decrease in the forest’s error rate. | + | Despite this trade-off, Random Forests balance these dynamics by optimising m to minimise the ensemble error. Generally, a moderate reduction in m lowers correlation without significantly compromising the strength of each tree, thus leading to an overall decrease in the forest's error rate. |
- | **Implications for the Forest Error Rate:** The forest error rate in a Random Forest model is influenced by both the correlation among the trees and the strength of each individual | + | **Implications for the Forest Error Rate:** The forest error rate in a Random Forest model is influenced by the correlation among the trees and the strength of each tree. Specifically: |
- | * Increasing correlation among trees typically increases the error rate, as it reduces the ensemble’s ability to correct individual trees' errors. | + | * Increasing correlation among trees typically increases the error rate, as it reduces the ensemble's ability to correct individual trees' errors. |
* Increasing the strength of each tree (i.e., reducing its error rate) generally decreases the forest error rate, as each tree becomes a more reliable classifier. | * Increasing the strength of each tree (i.e., reducing its error rate) generally decreases the forest error rate, as each tree becomes a more reliable classifier. | ||
Consequently, | Consequently, |