|
- Simple and interpretable
- Fast to train
|
- Assumes linear boundaries
- Not suitable for complex relationships
|
- Credit approval
- Medical diagnosis
|
|
|
- Intuitive
- Can model non-linear relationships
|
- Prone to overfitting
- Sensitive to small changes in data
|
- Customer segmentation
- Loan default prediction
|
|
|
- Handles overfitting
- Can model complex relationships
|
- Slower to train and predict
- Black box model
|
- Fraud detection
- Stock price movement prediction
|
|
|
- Effective in high dimensional spaces
- Works well with clear margin of separation
|
- Sensitive to kernel choice
- Slow on large datasets
|
- Image classification
- Handwriting recognition
|
|
|
- Simple and intuitive
- No training phase
|
- Slow during query phase
- Sensitive to irrelevant features and scale
|
- Product recommendation
- Document classification
|
|
|
- Capable of approximating complex functions
- Flexible architecture
Trainable with backpropagation
|
- Can require a large number of parameters
- Prone to overfitting on small data
Training can be slow
|
- Pattern recognition
- Basic image classification
- Function approximation
|
|
|
- Can model highly complex relationships
- Excels with vast amounts of data
State-of-the-art results in many domains
|
- Requires a lot of data
Computationally intensive
- Interpretability challenges
|
- Advanced image and speech recognition
- Machine translation
- Game playing (like AlphaGo)
|
|
|
- Fast
- Works well with large feature sets
|
- Assumes feature independence
- Not suitable for numerical input features
|
- Spam detection
- Sentiment analysis
|
|
|
- High performance
- Handles non-linear relationships
|
- Prone to overfitting if not tuned
- Slow to train
|
- Web search ranking
- Ecology predictions
|
|
|
- Transparent and explainable
- Easily updated and modified
|
- Manual rule creation can be tedious
- May not capture complex relationships
|
- Expert systems
- Business rule enforcement
|
|
|
- Reduces variance
- Parallelizable
|
|
- Random Forest is a popular example
|
|
|
- Reduces bias
- Combines weak learners
|
- Sensitive to noisy data and outliers
|
- AdaBoost
- Gradient Boosting
|
|
|
- Scalable and efficient
- Regularization
|
- Requires careful tuning
- Can overfit if not used correctly
|
- Competitions on Kaggle
- Retail prediction
|
|
|
- Dimensionality reduction
- Simple and interpretable
|
- Assumes Gaussian distributed data and equal class covariances
|
- Face recognition
- Marketing segmentation
|
|
|
- Prevents overfitting
- Handles collinearity
|
- Requires parameter tuning
- May result in loss of interpretability
|
- Ridge and Lasso regression
|
|
|
- Combines multiple models
- Can improve accuracy
|
- Increases model complexity
- Risk of overfitting if base models are correlated
|
- Meta-modeling
- Kaggle competitions
|
|