51 lines
1.6 KiB
Python
Vendored
51 lines
1.6 KiB
Python
Vendored
# -*- coding: utf-8 -*-
|
|
|
|
from fla.layers import (ABCAttention, BasedLinearAttention, DeltaNet,
|
|
GatedLinearAttention, HGRN2Attention, LinearAttention,
|
|
MultiScaleRetention, ReBasedLinearAttention)
|
|
from fla.models import (ABCForCausalLM, ABCModel, DeltaNetForCausalLM,
|
|
DeltaNetModel, GLAForCausalLM, GLAModel,
|
|
HGRN2ForCausalLM, HGRN2Model, HGRNForCausalLM,
|
|
HGRNModel, LinearAttentionForCausalLM,
|
|
LinearAttentionModel, RetNetForCausalLM, RetNetModel,
|
|
RWKV6ForCausalLM, RWKV6Model, TransformerForCausalLM,
|
|
TransformerModel)
|
|
from fla.ops import (chunk_gla, chunk_retention, fused_chunk_based,
|
|
fused_chunk_gla, fused_chunk_retention)
|
|
|
|
__all__ = [
|
|
'ABCAttention',
|
|
'BasedLinearAttention',
|
|
'DeltaNet',
|
|
'HGRN2Attention',
|
|
'GatedLinearAttention',
|
|
'LinearAttention',
|
|
'MultiScaleRetention',
|
|
'ReBasedLinearAttention',
|
|
'ABCForCausalLM',
|
|
'ABCModel',
|
|
'DeltaNetForCausalLM',
|
|
'DeltaNetModel',
|
|
'HGRNForCausalLM',
|
|
'HGRNModel',
|
|
'HGRN2ForCausalLM',
|
|
'HGRN2Model',
|
|
'GLAForCausalLM',
|
|
'GLAModel',
|
|
'LinearAttentionForCausalLM',
|
|
'LinearAttentionModel',
|
|
'RetNetForCausalLM',
|
|
'RetNetModel',
|
|
'RWKV6ForCausalLM',
|
|
'RWKV6Model',
|
|
'TransformerForCausalLM',
|
|
'TransformerModel',
|
|
'chunk_gla',
|
|
'chunk_retention',
|
|
'fused_chunk_based',
|
|
'fused_chunk_gla',
|
|
'fused_chunk_retention'
|
|
]
|
|
|
|
__version__ = '0.1'
|