Category: Artificial Intelligence

Activation Functions

Activation functions are mathematical formulas used in neural networks to decide whether a neuron should be activated or not. They help the network learn complex patterns by introducing non-linearity, allowing it to solve more complicated problems than a simple linear system could handle. Without activation functions, neural networks would not be able to model tasks…

Feature Attribution

Feature attribution is a method used in machine learning to determine how much each input feature contributes to a model’s prediction. It helps explain which factors are most important for the model’s decisions, making complex models more transparent. By understanding feature attribution, users can trust and interpret the outcomes of machine learning systems more easily.

Capsule Networks

Capsule Networks are a type of artificial neural network designed to better capture spatial relationships and hierarchies in data, such as images. Unlike traditional neural networks, capsules group neurons together to represent different properties of an object, like its position and orientation. This structure helps the network understand the whole object and its parts, making…

Residual Connections

Residual connections are a technique used in deep neural networks where the input to a layer is added to its output. This helps the network learn more effectively, especially as it becomes deeper. By allowing information to skip layers, residual connections make it easier for the network to avoid problems like vanishing gradients, which can…

Neural Tangent Kernel

The Neural Tangent Kernel (NTK) is a mathematical tool used to study and predict how very large neural networks learn. It simplifies the behaviour of neural networks by treating them like a type of kernel method, which is a well-understood class of machine learning models. Using the NTK, researchers can analyse training dynamics and generalisation…

LoRA Fine-Tuning

LoRA Fine-Tuning is a method used to adjust large pre-trained artificial intelligence models, such as language models, with less computing power and memory. Instead of changing all the model’s weights, LoRA adds small, trainable layers that adapt the model for new tasks. This approach makes it faster and cheaper to customise models for specific needs…