Subhaditya's KB

Home

❯

KB

❯

AI

❯

Machine Learning

❯

Models

❯

Multiplicative Attention

Multiplicative Attention

Sep 18, 20241 min read

  • architecture

Multiplicative Attention

  • fatt​(hi​,sj​)=hiT​Wa​sj​
  • Since Additive Attention performs better for scale, use a factor Scaled Dot Product Attention

Graph View

Backlinks

  • Dot Product Attention
  • _Index_of_Models
  • architecture

Created with Quartz v4.3.1 © 2025

  • GitHub