Category: Embeddings & Representations

Capsule Networks

Capsule Networks are a type of artificial neural network designed to better capture spatial relationships and hierarchies in data, such as images. Unlike traditional neural networks, capsules group neurons together to represent different properties of an object, like its position and orientation. This structure helps the network understand the whole object and its parts, making…

Tokenisation Strategies

Tokenisation strategies are methods used to split text into smaller pieces called tokens, which can be words, characters, or subwords. These strategies help computers process and understand language by breaking it down into more manageable parts. The choice of strategy can affect how well a computer model understands and generates text, as different languages and…

Contrastive Learning

Contrastive learning is a machine learning technique that teaches models to recognise similarities and differences between pairs or groups of data. It does this by pulling similar items closer together in a feature space and pushing dissimilar items further apart. This approach helps the model learn more useful and meaningful representations of data, even when…

Latent Space

Latent space refers to a mathematical space where complex data like images, sounds, or texts are represented as simpler numerical values. These values capture the essential features or patterns of the data, making it easier for computers to process and analyse. In machine learning, models often use latent space to find similarities, generate new examples,…