MultiHeadAttention ================== Computes Multi-Head Attention. **Abstract Signature:** ``MultiHeadAttention(embed_dim: int, num_heads: int)`` .. raw:: html

PyTorch

API: torch.nn.MultiheadAttention
Strategy: Direct Mapping

Keras

API: keras.layers.MultiHeadAttention
Strategy: Direct Mapping

TensorFlow

API: tf.keras.layers.MultiHeadAttention
Strategy: Direct Mapping

Apple MLX

API: mlx.nn.MultiHeadAttention
Strategy: Direct Mapping

Flax NNX

API: flax.nnx.MultiHeadAttention
Strategy: Direct Mapping

PaxML / Praxis

API: praxis.layers.MultiHeadAttention
Strategy: Direct Mapping