What are Query, Key, and Value in the Transformer Architecture and Why Are They Used? | by Ebrahim Pichka | Oct, 2023

An analysis of the intuition behind the notion of Key, Query, and Value in architecture and why is it used.

Ebrahim Pichka
Towards Data Science
Image by author — generated by Midjourney

Recent years have seen the Transformer architecture make waves in the field of (NLP), achieving results in a variety of tasks including machine translation, language modeling, and text summarization, as well as other domains of AI i.e. , Speech, RL, etc.

Vaswani et al. (2017), first introduced the transformer in their paper “Attention Is All You Need”, in which they used the mechanism without incorporating recurrent connections while the model can focus selectively on specific portions of input sequences.

The Transformer model architecture — Image from the Vaswani et al. (2017) paper (Source: arXiv:1706.03762v7)

In particular, previous sequence , such as recurrent -decoder models, were in their ability to capture long-term dependencies and parallel computations. In fact, right before the Transformers paper came out in 2017, state-of-the-art performance in most NLP tasks was obtained by using RNNs with an attention mechanism on top, so attention kind of existed before transformers. By introducing the multi-head attention mechanism on its own, and dropping the RNN part, the transformer architecture resolves these issues by allowing multiple independent attention mechanisms.

In this post, we will go over one of the details of this architecture, namely the Query, Key, and Values, and try to make sense of the intuition used behind this part.

Note that this post assumes you are already familiar with some basic concepts in NLP and such as embeddings, Linear (dense) layers, and in general how a simple neural network works.

First, let’s start understanding what the attention mechanism is trying to achieve. And for the sake of simplicity, let’s start with a simple case of sequential to understand what problem exactly…

Source link