Deep Learning Concepts
Expert-defined terms from the Professional Certificate in Artificial Intelligence for K-12 Educators course at HealthCareStudies (An LSPM brand). Free to read, free to share, paired with a globally recognised certification pathway.
Activation Function #
A function that introduces non-linearity into a neural network by transforming the input signal into an output signal. Common activation functions include ReLU (Rectified Linear Unit), Sigmoid, and Tanh.
Artificial Neural Network (ANN) #
A computational model inspired by the human brain's biological neural networks, consisting of interconnected nodes (neurons) organized in layers. ANNs are used in deep learning for various tasks such as image recognition and natural language processing.
Backpropagation #
A supervised learning algorithm used to train neural networks by adjusting the weights of connections between neurons in the network. Backpropagation calculates the gradient of the loss function with respect to the weights, allowing for the optimization of the network's performance.
Batch #
A set of training examples used in one iteration of the training process. Batching training data helps improve computational efficiency and generalization performance in deep learning models.
Convolutional Neural Network (CNN) #
A type of deep neural network commonly used for image recognition and computer vision tasks. CNNs consist of convolutional layers that apply filters to extract features from input images, followed by pooling layers for downsampling.
Deep Learning #
A subset of machine learning that uses neural networks with multiple layers (deep neural networks) to learn complex patterns and representations from data. Deep learning has been successful in various domains, including computer vision, natural language processing, and speech recognition.
Epoch #
In deep learning, an epoch refers to one complete pass through the entire training dataset during the training process. Training deep learning models typically involves multiple epochs to improve model performance.
Feature Extraction #
The process of automatically discovering and selecting relevant features from raw data to improve the performance of machine learning models. Deep learning models use feature extraction to learn hierarchical representations of data.
Gradient Descent #
An optimization algorithm used to minimize the loss function of a neural network by iteratively adjusting the weights and biases of the network in the direction of the steepest descent of the gradient.
Hyperparameter #
Parameters that are set before the training process begins and affect the learning process of a deep learning model. Examples of hyperparameters include learning rate, batch size, and number of layers in a neural network.
Image Classification #
A computer vision task that involves categorizing images into predefined classes or labels. Deep learning models, such as convolutional neural networks, are commonly used for image classification tasks.
Loss Function #
A function that calculates the difference between the predicted output of a neural network and the actual target output during the training process. The goal of training is to minimize the loss function to improve the model's performance.
Neural Network #
A computational model inspired by the structure and function of the human brain's neural networks. Neural networks consist of interconnected nodes (neurons) organized in layers that process input data to produce output predictions.
Overfitting #
A common problem in deep learning where a model performs well on the training data but fails to generalize to unseen data. Overfitting occurs when a model learns noise or irrelevant patterns in the training data.
Pooling Layer #
A type of layer in a convolutional neural network that reduces the spatial dimensions of feature maps produced by convolutional layers. Pooling layers help extract dominant features from the input data while reducing computational complexity.
Recurrent Neural Network (RNN) #
A type of neural network architecture designed to process sequential data with time dependencies. RNNs have recurrent connections that allow information to persist over time, making them suitable for tasks such as speech recognition and natural language processing.
Supervised Learning #
A machine learning paradigm where the model is trained on labeled data, meaning the input data is paired with corresponding output labels. Supervised learning is commonly used in deep learning for tasks such as classification and regression.
Transfer Learning #
A technique in deep learning where a pre-trained model is used as a starting point for a new task. Transfer learning leverages the knowledge learned from one task to improve the performance of a model on a related task with limited labeled data.
Unsupervised Learning #
A machine learning paradigm where the model learns patterns and relationships from unlabeled data without explicit supervision. Unsupervised learning is used in deep learning for tasks such as clustering and dimensionality reduction.
Vanishing Gradient Problem #
A challenge in training deep neural networks where the gradients of the loss function become extremely small as they propagate backward through the network layers. The vanishing gradient problem can hinder the training of deep learning models with many layers.