
Information
- Bagging (Bootstrap Aggregating) is a machine learning technique that combines multiple models trained on different subsets of the training data.
 - Bagging is often used to reduce the variance and improve the stability of the predictions.
 - Bagging samples the training data with replacement to create multiple bootstrap samples, each of which has the same size as the original dataset.
 - Bagging trains a separate model on each bootstrap sample, using the same model type and hyperparameters.
 - The final prediction is a combination of the predictions of all the models, typically by averaging the predictions for regression problems or using voting for classification problems.
 - Bagging can be used for both classification and regression problems.
 - Random Forest is a popular bagging algorithm for decision trees, which randomly selects a subset of features at each split to decorrelate the trees and improve the generalization.
 - Random Forest can handle high-dimensional data and non-linear relationships between the features and the target variable.
 - Random Forest can also estimate the importance of each feature based on the reduction in impurity, and thus provide insights into the underlying data.
 - Bagging algorithms have achieved state-of-the-art performance on many machine learning benchmarks and have been widely used in various applications, such as bioinformatics, finance, and e-commerce.
 
'Zettelkasten > Terminology Information' 카테고리의 다른 글
| SELU (Scaled Exponential Linear Unit) (0) | 2023.02.24 | 
|---|---|
| Gradient vanishing (0) | 2023.02.24 | 
| Long-term dependency (0) | 2023.02.23 | 
| Boosting (0) | 2023.02.22 | 
| MSE (Mean Squared Error) (0) | 2023.02.22 | 
			
			
				
			
댓글