Get in Touch

Course Outline

Introduction

This section offers a broad overview of when to apply machine learning, key considerations, and the underlying concepts, including advantages and limitations. Topics include data types (structured, unstructured, static, streamed), data validity and volume, data-driven versus user-driven analytics, and a comparison of statistical models against machine learning models. It also covers the challenges of unsupervised learning, the bias-variance trade-off, iteration and evaluation processes, cross-validation techniques, and the distinctions among supervised, unsupervised, and reinforcement learning.

MAJOR TOPICS

1. Understanding naive Bayes

  • Core concepts of Bayesian methods
  • Probability theory
  • Joint probability
  • Conditional probability via Bayes' theorem
  • The naive Bayes algorithm
  • Naive Bayes classification
  • The Laplace estimator
  • Applying numeric features with naive Bayes

2. Understanding decision trees

  • Divide and conquer strategy
  • The C5.0 decision tree algorithm
  • Selecting the optimal split
  • Pruning the decision tree

3. Understanding neural networks

  • Evolution from biological to artificial neurons
  • Activation functions
  • Network topology
  • Determining the number of layers
  • Direction of information flow
  • Node count per layer
  • Training neural networks using backpropagation
  • Deep Learning

4. Understanding Support Vector Machines

  • Classification using hyperplanes
  • Maximizing the margin
  • Handling linearly separable data
  • Handling non-linearly separable data
  • Utilizing kernels for non-linear spaces

5. Understanding clustering

  • Clustering as a machine learning task
  • The k-means clustering algorithm
  • Using distance metrics to assign and update clusters
  • Choosing the appropriate number of clusters

6. Measuring performance for classification

  • Working with classification prediction data
  • An in-depth look at confusion matrices
  • Using confusion matrices to assess performance
  • Beyond accuracy – other performance metrics
  • The kappa statistic
  • Sensitivity and specificity
  • Precision and recall
  • The F-measure
  • Visualizing performance trade-offs
  • ROC curves
  • Estimating future performance
  • The holdout method
  • Cross-validation
  • Bootstrap sampling

7. Tuning stock models for better performance

  • Using caret for automated parameter tuning
  • Creating a simple tuned model
  • Customizing the tuning process
  • Enhancing model performance with meta-learning
  • Understanding ensembles
  • Bagging
  • Boosting
  • Random forests
  • Training random forests
  • Evaluating random forest performance

MINOR TOPICS

8. Understanding classification using the nearest neighbors

  • The kNN algorithm
  • Calculating distance
  • Choosing an appropriate k
  • Preparing data for use with kNN
  • Why is the kNN algorithm lazy?

9. Understanding classification rules

  • Separate and conquer approach
  • The One Rule algorithm
  • The RIPPER algorithm
  • Deriving rules from decision trees

10. Understanding regression

  • Simple linear regression
  • Ordinary least squares estimation
  • Correlations
  • Multiple linear regression

11. Understanding regression trees and model trees

  • Incorporating regression into trees

12. Understanding association rules

  • The Apriori algorithm for association rule learning
  • Measuring rule interest – support and confidence
  • Building a set of rules with the Apriori principle

Extras

  • Spark/PySpark/MLlib and Multi-armed bandits

Requirements

Knowledge of Python

 21 Hours

Number of participants


Price per participant

Testimonials (7)

Upcoming Courses

Related Categories