Navigating the Attention Landscape: MHA, MQA, and GQA Decoded

Shobhit Agarwal
5 min readJan 4, 2024

--

Attention mechanisms are the driving force behind many of today’s cutting-edge Large Language Models. They allow these models to focus on relevant parts of an input sequence, like a sentence or document, and extract meaning with astonishing accuracy. But with different flavours of attention popping up, things can get confusing. Today, we’ll explore three key players:

  1. Multi-Head Attention (MHA)
  2. Multi-Query Attention (MQA)
  3. Grouped-Query Attention (GQA)

Whether you’re a complete beginner or someone with a basic understanding of NLP, I tried to make this article designed in a way to provide clear explanations related to MHA, MQA and GQA.

The Attention Spotlight:

  • Attention mechanisms focus on relevant parts of an input sequence like sentences or documents.
  • Imagine a magician pulling a rabbit out of a hat — attention mechanisms selectively “reveal” important information.
  • Attention plays a crucial role in tasks like understanding sentiment, translating languages, and answering questions.
  1. Multi-Head Attention: The Star Player
  • Point 1: MHA uses multiple “heads” to attend to different aspects of the input simultaneously, like a multi-tasking detective.
  • Point 2: Think of it as reading a newspaper with multiple headlines — MHA captures various nuances of the text.
  • Point 3: MHA provides high-quality results but can be computationally expensive due to parallel processing.

2. Multi-Query Attention: The Speedster

  • Point 1: MQA uses only one head for faster processing, like a sprinter racing through the input.
  • Point 2: Imagine skimming a document for keywords — MQA quickly grasps the main gist of the content.
  • Point 3: While efficient, MQA might miss important details compared to MHA’s in-depth analysis.

3. Grouped-Query Attention: The Rising Star

  • Point 1: GQA balances speed and detail by grouping queries and focusing attention on those groups.
  • Point 2: Think of it as dividing a team into smaller units to tackle specific tasks, yet collaborating overall.
  • Point 3: GQA offers faster processing than MHA while retaining some detail through focused attention within groups.

There are many open-source LLMs available in the market that work on the GQA mechanism.

Figure: Llama2–70B model uses GQA

Let’s try to understand Attention Heads with the following Analogy

Figure: Imagine you’re at a bustling party with different groups of people talking (the keys and values). You have multiple friends you want to catch up with (the queries).

Imagine you’re at a bustling party with different groups of people talking (the keys and values). You have multiple friends you want to catch up with (the queries).

Multi-Head Attention (MHA): You listen to each friend individually, focusing on each conversation one by one. This gives you all the details but takes a long time!

Multi-Query Attention (MQA): You shout over everyone at once, trying to catch snippets of every conversation. This lets you know what topics are floating around quickly, but you miss important details about each individual conversation.

Grouped-Query Attention (GQA): You divide your friends into smaller groups. You listen to each group as a whole, catching the main points of their conversations. This is faster than MHA but still gives you more details than MQA because you’re focused on smaller groups instead of everyone at once.

So, GQA is like a compromise between MHA and MQA, giving you a balance between speed and detail. It’s a way for large language models to understand multiple things at once without getting overwhelmed, like you at the party!

Remember, GQA is like having group conversations instead of individual ones — you get the gist of things quickly while still focusing on smaller chunks of information. This makes it faster and more efficient for models to process complex inputs.

Hopefully, this analogy will help you recall MHA, MQA and GQA easily in the future!

Now, if you are still with me, let me bring some technicalities involved in each attention mechanism.

Technical breakdown of MHA, MQA and GQA

Source: https://arxiv.org/pdf/2305.13245.pdf

Multi-Head Attention (MHA):

  • Uses H separate “heads” for both queries, keys, and values.
  • Each head attends to different aspects of the input independently.
  • Provides high quality but is computationally expensive due to many independent calculations.

Multi-Query Attention (MQA):

  • Uses only one head for all queries, keys, and values.
  • Performs a single attention calculation over the entire input.
  • Much faster than MHA but can produce lower-quality results due to a lack of individual focus.

Grouped-Query Attention (GQA):

  • Acts as a middle ground between MHA and MQA.
  • Divide the H query heads into G groups.
  • Each group shares a single key and value head.
  • Calculates attention within each group, allowing some differentiation.
  • Offers a trade-off:
  • Faster than MHA due to fewer heads per group.
  • Higher quality than MQA due to focused attention within groups.

Here’s an example:

Imagine you’re analyzing a document with three main topics: technology, politics, and sports.

  • MHA: You have three separate teams (heads) analyzing each topic independently. They get a detailed understanding of each topic but it takes longer.
  • MQA: You have one big team analyzing the entire document at once. They get a quick overview but might miss crucial details within each topic.
  • GQA: You divide the analysis into smaller teams (groups) — one for technology, one for politics, and one for sports. Each team still has individual focus but can collaborate when needed. This provides a quicker understanding while preserving some detail.

GQA uses this group-based approach to achieve an efficient balance between speed and quality, making it a valuable tool for large language models.

I hope this technical explanation solidifies your understanding of each.

Now, let’s take a look at the future outlook.

Future Outlook

  • Research is exploring new attention mechanisms like hybrid approaches and dynamic routing.
  • Attention is expected to play an even bigger role in future NLP tasks like language generation and dialogue systems.

The future of NLP is bright, and with advancements like GQA, the possibilities are endless. What exciting applications do you envision for these attention mechanisms?

Share your thoughts in the comments!

--

--

Shobhit Agarwal

Data Scientist | Specializing in ML & DL | Scaling LLMs & Gen AI | Python expertise | GCP Certified