AIbase
Product LibraryTool Navigation

memory-efficient-attention-pytorch

Public

Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n2) Memory"

Creat2022-03-03T14:26:31
Update2025-03-26T17:17:23
375
Stars
0
Stars Increase