modules/SwissArmyTransformer/sat/model/attention/__init__.py (
1
lines of code) (
raw
):
from .memory_efficient_attention import TransposedMemoryEfficientAttentionMixin, MemoryEfficientAttentionMixin