Category: Model Optimisation Techniques

Gradient Clipping

Gradient clipping is a technique used in training machine learning models to prevent the gradients from becoming too large during backpropagation. Large gradients can cause unstable training and make the model’s learning process unreliable. By setting a maximum threshold, any gradients exceeding this value are scaled down, helping to keep the learning process steady and…

Quantisation-Aware Training

Quantisation-Aware Training is a method used to prepare machine learning models for running efficiently on devices with limited computing power, such as smartphones or embedded systems. It teaches the model to handle the reduced precision of numbers, which happens when large models are made smaller by using fewer bits to represent data. This approach helps…

Activation Functions

Activation functions are mathematical formulas used in neural networks to decide whether a neuron should be activated or not. They help the network learn complex patterns by introducing non-linearity, allowing it to solve more complicated problems than a simple linear system could handle. Without activation functions, neural networks would not be able to model tasks…

Residual Connections

Residual connections are a technique used in deep neural networks where the input to a layer is added to its output. This helps the network learn more effectively, especially as it becomes deeper. By allowing information to skip layers, residual connections make it easier for the network to avoid problems like vanishing gradients, which can…

Gradient Accumulation

Gradient accumulation is a technique used in training neural networks where gradients from several smaller batches are summed before updating the model’s weights. This allows the effective batch size to be larger than what would normally fit in memory. It is especially useful when hardware limitations prevent the use of large batch sizes during training.

Parameter-Efficient Fine-Tuning

Parameter-efficient fine-tuning is a machine learning technique that adapts large pre-trained models to new tasks or data by modifying only a small portion of their internal parameters. Instead of retraining the entire model, this approach updates selected components, which makes the process faster and less resource-intensive. This method is especially useful when working with very…