Attention.jl

Welcome to the documentation for Attention.jl.

This package provides modular and extensible attention mechanisms for deep learning models in Julia.

Installation

using Pkg
Pkg.add(url="https://github.com/mashu/Attention.jl.git")

Main Features

  • Modular attention mechanism interface
  • DotProductAttention, NNlibAttention, LinearAttention (currently non-causal and global; causal support planned), MultiHeadAttention
  • Utilities like make_causal_mask
  • Support for custom Q/K transformations (e.g., RoPE)

Check out the API Reference for detailed information on the functions and types provided.