Fundamentals of Machine Learning
Expert-defined terms from the Professional Certificate in AI for Operations Management course at HealthCareStudies (An LSPM brand). Free to read, free to share, paired with a globally recognised certification pathway.
Active Learning is a subfield of machine learning that involves actively… #
Related terms include semi-supervised learning, transfer learning, and reinforcement learning. Active learning is particularly useful in applications where labeling data is expensive or time-consuming, such as in medical imaging or text classification. For example, in a medical diagnosis task, active learning can be used to select the most informative images to be labeled by a doctor, reducing the amount of time and effort required to achieve accurate diagnosis.
Adversarial Attack refers to a type of attack on a machine learning model… #
Related terms include adversarial training, robustness, and security. Adversarial attacks can be used to evaluate the robustness of a machine learning model, and to develop more robust models that are resistant to such attacks. For example, in a self-driving car application, an adversarial attack could be used to manipulate the input images to cause the model to misclassify a stop sign as a speed limit sign.
Autoencoder is a type of neural network that is trained to copy it… #
Related terms include encoder, decoder, and bottleneck. Autoencoders are useful in applications where the input data is high-dimensional and needs to be reduced to a lower dimensionality for easier processing or analysis. For example, in an image compression task, an autoencoder can be used to reduce the dimensionality of the input images while preserving the most important features.
Backpropagation is an algorithm used to train neural networks, which invo… #
Related terms include optimization, gradient descent, and chain rule. Backpropagation is a key component of many machine learning algorithms, including deep learning models. For example, in a language modeling task, backpropagation can be used to train a neural network to predict the next word in a sentence based on the context.
Batch Normalization is a technique used to normalize the input data to a… #
Related terms include normalization, standardization, and regularization. Batch normalization is useful in applications where the input data has a large range of values, and can help to improve the stability and speed of training. For example, in an image classification task, batch normalization can be used to normalize the input images to have a mean of zero and a variance of one, which can help to improve the accuracy of the model.
Bayes' Theorem is a mathematical formula used to update the probabilit… #
Related terms include Bayesian inference, probability theory, and statistics. Bayes' theorem is useful in applications where there is uncertainty or noise in the data, and can help to make more informed decisions. For example, in a medical diagnosis task, Bayes' theorem can be used to update the probability of a disease based on new evidence, such as test results or symptoms.
Bias #
Variance Tradeoff refers to the tradeoff between the bias and variance of a machine learning model, where a model with low bias may have high variance, and a model with low variance may have high bias. Related terms include overfitting, underfitting, and regularization. The bias-variance tradeoff is a fundamental concept in machine learning, and is used to evaluate the performance of a model. For example, in a regression task, a model with low bias may have high variance, resulting in overfitting to the training data, while a model with low variance may have high bias, resulting in underfitting to the training data.
Clustering is a type of unsupervised learning algorithm that involves gro… #
Related terms include k-means, hierarchical clustering, and density-based clustering. Clustering is useful in applications where there is no labeled data, and can help to identify patterns and structures in the data. For example, in a customer segmentation task, clustering can be used to group similar customers into clusters based on their behavior and demographics.
Convolutional Neural Network (CNN) is a type of neural network that is de… #
Related terms include convolution, pooling, and fully connected. CNNs are useful in applications where the input data has a spatial hierarchy, and can help to extract features and patterns from the data. For example, in an image classification task, a CNN can be used to extract features from the input images and classify them into different categories.
Decision Tree is a type of machine learning model that involves using a <… #
Related terms include random forest, gradient boosting, and splitting. Decision trees are useful in applications where the input data has a complex structure, and can help to identify patterns and relationships between the variables. For example, in a credit risk assessment task, a decision tree can be used to classify customers into different risk categories based on their credit history and behavior.
Deep Learning is a subfield of machine learning that involves using neura… #
Related terms include neural network, convolutional neural network, and recurrent neural network. Deep learning is useful in applications where the input data has a complex structure, and can help to extract features and patterns from the data. For example, in a self-driving car application, deep learning can be used to extract features from the input images and sensors to detect and respond to objects in the environment.
Dimensionality Reduction is a technique used to reduce the number of f… #
Related terms include principal component analysis, t-SNE, and autoencoder. Dimensionality reduction is useful in applications where the input data has a high dimensionality and needs to be reduced to a lower dimensionality for easier processing or analysis. For example, in a gene expression analysis task, dimensionality reduction can be used to reduce the number of genes in the dataset to a smaller set of features that are most relevant to the condition being studied.
Ensemble Learning is a technique used to combine the predictions of multi… #
Related terms include bagging, boosting, and stacking. Ensemble learning is useful in applications where a single model is not sufficient to capture the complexity of the data, and can help to improve the accuracy and robustness of the predictions. For example, in a credit risk assessment task, ensemble learning can be used to combine the predictions of multiple models to produce a more accurate risk score for each customer.
Feature Engineering is the process of selecting and transforming raw data… #
Related terms include feature extraction, feature selection, and data preprocessing. Feature engineering is a critical step in the machine learning pipeline, and can help to improve the accuracy and efficiency of the models. For example, in a text classification task, feature engineering can be used to extract features from the input text data, such as word frequencies and sentiment analysis.
Gradient Boosting is a type of ensemble learning algorithm that involves… #
Related terms include boosting, bagging, and random forest. Gradient boosting is useful in applications where a single model is not sufficient to capture the complexity of the data, and can help to improve the accuracy and robustness of the predictions. For example, in a credit risk assessment task, gradient boosting can be used to combine the predictions of multiple models to produce a more accurate risk score for each customer.
Hyperparameter Tuning is the process of selecting the optimal hyperpar… #
Related terms include grid search, random search, and Bayesian optimization. Hyperparameter tuning is a critical step in the machine learning pipeline, and can help to improve the accuracy and efficiency of the models. For example, in a classification task, hyperparameter tuning can be used to select the optimal hyperparameters for a logistic regression model, such as the regularization parameter and the learning rate.
K-Means Clustering is a type of unsupervised learning algorithm that invo… #
Related terms include clustering, hierarchical clustering, and density-based clustering. K-means clustering is useful in applications where there is no labeled data, and can help to identify patterns and structures in the data. For example, in a customer segmentation task, k-means clustering can be used to group similar customers into k clusters based on their behavior and demographics.
Linear Regression is a type of supervised learning algorithm that involve… #
Related terms include regression, least squares, and ordinary least squares. Linear regression is useful in applications where the relationship between the variables is linear, and can help to identify the strength and direction of the relationships. For example, in a stock market prediction task, linear regression can be used to model the relationship between the stock price and various economic indicators.
Logistic Regression is a type of supervised learning algorithm that invol… #
Related terms include classification, logit, and odds ratio. Logistic regression is useful in applications where the outcome is binary, and can help to identify the strength and direction of the relationships between the variables. For example, in a credit risk assessment task, logistic regression can be used to model the probability of default based on various credit variables.
Natural Language Processing (NLP) is a subfield of artificial intelligenc… #
Related terms include text mining, information retrieval, and machine translation. NLP is useful in applications where the input data is in the form of text, and can help to extract meaning and insights from the data. For example, in a sentiment analysis task, NLP can be used to analyze the text data to determine the sentiment of the customers towards a particular product or service.
Neural Network is a type of machine learning model that is inspired by th… #
Related terms include deep learning, convolutional neural network, and recurrent neural network. Neural networks are useful in applications where the input data has a complex structure, and can help to extract features and patterns from the data. For example, in a self-driving car application, neural networks can be used to extract features from the input images and sensors to detect and respond to objects in the environment.
Overfitting is a problem that occurs when a machine learning model is too… #
Related terms include underfitting, regression, and regularization. Overfitting is a common problem in machine learning, and can be addressed using techniques such as regularization, early stopping, and ensemble learning. For example, in a classification task, overfitting can occur when the model is too complex and fits the noise in the training data, resulting in poor accuracy on new, unseen data.
Principal Component Analysis (PCA) is a technique used to reduce the d… #
Related terms include dimensionality reduction, feature selection, and data preprocessing. PCA is useful in applications where the input data has a high dimensionality and needs to be reduced to a lower dimensionality for easier processing or analysis. For example, in a gene expression analysis task, PCA can be used to reduce the number of genes in the dataset to a smaller set of features that are most relevant to the condition being studied.
Random Forest is a type of ensemble learning algorithm that involves comb… #
Related terms include bagging, boosting, and gradient boosting. Random forest is useful in applications where a single model is not sufficient to capture the complexity of the data, and can help to improve the accuracy and robustness of the predictions. For example, in a credit risk assessment task, random forest can be used to combine the predictions of multiple decision trees to produce a more accurate risk score for each customer.
Recurrent Neural Network (RNN) is a type of neural network that is design… #
Related terms include long short-term memory, gated recurrent unit, and sequence-to-sequence modeling. RNNs are useful in applications where the input data has a temporal structure, and can help to extract features and patterns from the data. For example, in a language modeling task, RNNs can be used to predict the next word in a sentence based on the context.
Regularization is a technique used to prevent overfitting in machi… #
Related terms include dropout, early stopping, and ensemble learning. Regularization is useful in applications where the model is prone to overfitting, and can help to improve the generalization of the model to new, unseen data. For example, in a classification task, regularization can be used to prevent overfitting by adding a penalty term to the loss function.
Robustness is the ability of a machine learning model to perform well in… #
Related terms include regularization, early stopping, and ensemble learning. Robustness is a critical aspect of machine learning, and can help to improve the accuracy and reliability of the models. For example, in a classification task, robustness can be used to evaluate the performance of a model in the presence of noise or outliers in the data.
Supervised Learning is a type of machine learning that involves training… #
Related terms include unsupervised learning, semi-supervised learning, and reinforcement learning. Supervised learning is useful in applications where there is a clear definition of the target variable, and can help to make accurate predictions on new, unseen data! For example, in a credit risk assessment task, supervised learning can be used to train a model on labeled data to predict the risk of default for new customers.
Support Vector Machine (SVM) is a type of supervised learning algorithm t… #
Related terms include kernel trick, soft margin, and hard margin. SVMs are useful in applications where the classes are linearly separable, and can help to make accurate predictions on new, unseen data. For example, in a text classification task, SVMs can be used to classify text documents into different categories based on their content.
Unsupervised Learning is a type of machine learning that involves trainin… #
Related terms include supervised learning, semi-supervised learning, and reinforcement learning. Unsupervised learning is useful in applications where there is no labeled data, and can help to identify patterns and relationships in the data. For example, in a customer segmentation task, unsupervised learning can be used to group similar customers into clusters based on their behavior and demographics.
Validation is the process of evaluating the performance of a machine lear… #
Related terms include training, testing, and cross-validation. Validation is a critical step in the machine learning pipeline, and can help to evaluate the performance of a model and prevent overfitting. For example, in a classification task, validation can be used to evaluate the performance of a model on a holdout dataset to estimate its performance on new, unseen data.