MultiHeadAttentionΒΆ

Computes Multi-Head Attention.

Abstract Signature:

MultiHeadAttention(embed_dim: int, num_heads: int)

PyTorch

API: torch.nn.MultiheadAttention
Strategy: Direct Mapping

Keras

API: keras.layers.MultiHeadAttention
Strategy: Direct Mapping

TensorFlow

API: tf.keras.layers.MultiHeadAttention
Strategy: Direct Mapping

Apple MLX

API: mlx.nn.MultiHeadAttention
Strategy: Direct Mapping

Flax NNX

API: flax.nnx.MultiHeadAttention
Strategy: Direct Mapping

PaxML / Praxis

API: praxis.layers.MultiHeadAttention
Strategy: Direct Mapping