Machine Learning Algorithms

Machine Learning (ML) is a subfield of Artificial Intelligence (AI) that focuses on the development of algorithms that can learn patterns and make predictions from data. In the Professional Certificate in AI-Powered Business Analysis, you w…

Machine Learning Algorithms

Machine Learning (ML) is a subfield of Artificial Intelligence (AI) that focuses on the development of algorithms that can learn patterns and make predictions from data. In the Professional Certificate in AI-Powered Business Analysis, you will learn about various ML algorithms, their applications, and how to implement them in practice. In this explanation, we will cover key terms and vocabulary related to ML algorithms that will be useful for the course.

1. Machine Learning Algorithms

Machine learning algorithms can be broadly categorized into three types: supervised learning, unsupervised learning, and reinforcement learning.

Supervised learning is a type of ML algorithm where the model is trained on labeled data, i.e., data with known outputs. The model learns to map inputs to outputs based on the training data and can then make predictions on new, unseen data. Common supervised learning algorithms include linear regression, logistic regression, decision trees, and support vector machines (SVMs).

Unsupervised learning is a type of ML algorithm where the model is trained on unlabeled data, i.e., data without known outputs. The model learns to identify patterns and structure in the data without any explicit guidance. Common unsupervised learning algorithms include clustering algorithms (e.g., k-means, hierarchical clustering) and dimensionality reduction algorithms (e.g., principal component analysis (PCA), t-distributed stochastic neighbor embedding (t-SNE)).

Reinforcement learning is a type of ML algorithm where an agent learns to make decisions by interacting with an environment. The agent receives feedback in the form of rewards or penalties and learns to maximize the rewards over time. Reinforcement learning algorithms include Q-learning, deep Q-learning, and policy gradients.

2. Linear Regression

Linear regression is a simple supervised learning algorithm that models the relationship between a dependent variable (y) and one or more independent variables (x) using a linear function. The goal of linear regression is to find the best-fitting line or hyperplane that minimizes the sum of squared residuals. Linear regression can be used for regression tasks, where the output is a continuous variable.

3. Logistic Regression

Logistic regression is a supervised learning algorithm used for classification tasks, where the output is a categorical variable. Logistic regression models the probability of the output belonging to a certain class using the logistic function, which maps any real-valued number to a probability between 0 and 1.

4. Decision Trees

Decision trees are a type of supervised learning algorithm used for both regression and classification tasks. A decision tree is a hierarchical model that recursively partitions the data into subsets based on the values of the input features. The tree consists of nodes, where each node represents a decision based on a feature, and edges, which represent the outcome of the decision. The leaves of the tree represent the final predictions.

5. Support Vector Machines (SVMs)

SVMs are a type of supervised learning algorithm used for classification tasks. SVMs find the hyperplane that maximally separates the data points of different classes while minimizing the margin, i.e., the distance between the hyperplane and the closest data points. SVMs can also be used for regression tasks, in which case they are called support vector regression (SVR).

6. k-Means Clustering

k-means clustering is a type of unsupervised learning algorithm used for clustering tasks, where the goal is to group similar data points together. k-means clustering partitions the data into k clusters, where k is a user-defined parameter. The algorithm iteratively assigns each data point to the closest centroid and updates the centroids based on the mean of the data points in each cluster.

7. Principal Component Analysis (PCA)

PCA is a type of unsupervised learning algorithm used for dimensionality reduction, where the goal is to reduce the number of features while preserving as much information as possible. PCA identifies the principal components, which are linear combinations of the original features that capture the most variance in the data. The data can then be projected onto the principal components, reducing the number of features while retaining the most important information.

8. Overfitting and Underfitting

Overfitting and underfitting are common problems in machine learning, where the model does not generalize well to new, unseen data. Overfitting occurs when the model is too complex and fits the training data too closely, capturing the noise and irrelevant patterns in the data. Underfitting occurs when the model is too simple and cannot capture the underlying patterns in the data. To avoid overfitting and underfitting, it is important to use regularization techniques, such as L1 and L2 regularization, dropout, and early stopping.

9. Cross-Validation

Cross-validation is a technique used to evaluate the performance of a machine learning model and avoid overfitting. In cross-validation, the data is split into k folds, where k is a user-defined parameter. The model is trained on k-1 folds and tested on the remaining fold. This process is repeated k times, with a different fold used for testing each time. The performance of the model is then averaged over the k iterations.

10. Hyperparameter Tuning

Hyperparameter tuning is the process of finding the best set of hyperparameters for a machine learning model. Hyperparameters are parameters that are not learned from the data, such as the learning rate, regularization strength, and number of hidden layers in a neural network. Hyperparameter tuning can be done manually, using a grid search or random search, or using automated methods, such as Bayesian optimization.

11. Bias and Variance

Bias and variance are related to overfitting and underfitting and are important concepts in machine learning. Bias is the error introduced by assuming a simplified model, while variance is the error introduced by sensitivity to small fluctuations in the training data. A model with high bias is likely to underfit the data, while a model with high variance is likely to overfit the data. The goal is to find a balance between bias and variance, where the model has low error on both the training data and the test data.

12. Activation Functions

Activation functions are used in neural networks to introduce non-linearity and allow the model to learn complex patterns in the data. Common activation functions include the sigmoid function, the tanh function, and the rectified linear unit (ReLU) function. The choice of activation function can have a significant impact on the performance of the model.

13. Backpropagation

Backpropagation is an optimization algorithm used to train neural networks. The algorithm computes the gradient of the loss function with respect to the weights and biases of the network, allowing the model to learn the optimal parameters. Backpropagation is a form of gradient descent and is used to update the weights and biases of the network iteratively.

14. Convolutional Neural Networks (CNNs)

Convolutional neural networks (CNNs) are a type of neural network used for image and video analysis. CNNs consist of convolutional layers, pooling layers, and fully connected layers. The convolutional layers apply filters to the input data to extract features, while the pooling layers reduce the spatial dimensions of the data. The fully connected layers perform the final classification or regression task.

15. Recurrent Neural Networks (RNNs)

Recurrent neural networks (RNNs) are a type of neural network used for sequential data analysis, such as time series analysis and natural language processing. RNNs have a feedback loop that allows the model to incorporate information from previous time steps. Long short-term memory (LSTM) networks and gated recurrent unit (GRU) networks are variants of RNNs that can handle long-range dependencies in the data.

16. Transfer Learning

Transfer learning is a technique used in machine learning to leverage pre-trained models for new tasks. In transfer learning, the model is pre-trained on a large dataset and fine-tuned on a smaller dataset for the new task. Transfer learning can save time and resources and can lead to better performance, especially when the new dataset is small or noisy.

17. Explainable AI (XAI)

Explainable AI (XAI) is the practice of designing machine learning models that are transparent, interpretable, and understandable to humans. XAI is important in business analysis, where it is crucial to understand the decisions made by the model and to build trust with stakeholders. XAI techniques include feature importance, partial dependence plots, and local interpretable model-agnostic explanations (LIME).

18. Ethics and Fairness

Ethics and fairness are important considerations in machine learning, especially in business analysis, where the model may have a

Key takeaways

  • Machine Learning (ML) is a subfield of Artificial Intelligence (AI) that focuses on the development of algorithms that can learn patterns and make predictions from data.
  • Machine learning algorithms can be broadly categorized into three types: supervised learning, unsupervised learning, and reinforcement learning.
  • Common supervised learning algorithms include linear regression, logistic regression, decision trees, and support vector machines (SVMs).
  • Unsupervised learning is a type of ML algorithm where the model is trained on unlabeled data, i.
  • Reinforcement learning is a type of ML algorithm where an agent learns to make decisions by interacting with an environment.
  • Linear regression is a simple supervised learning algorithm that models the relationship between a dependent variable (y) and one or more independent variables (x) using a linear function.
  • Logistic regression models the probability of the output belonging to a certain class using the logistic function, which maps any real-valued number to a probability between 0 and 1.
May 2026 intake · open enrolment
from £90 GBP
Enrol