| # -*- coding: utf-8 -*- | |
| from fla.layers import ( | |
| ABCAttention, | |
| Attention, | |
| BasedLinearAttention, | |
| BitAttention, | |
| DeltaNet, | |
| GatedDeltaNet, | |
| GatedDeltaProduct, | |
| GatedLinearAttention, | |
| GatedSlotAttention, | |
| HGRN2Attention, | |
| HGRNAttention, | |
| LightNetAttention, | |
| LinearAttention, | |
| MultiScaleRetention, | |
| NativeSparseAttention, | |
| ReBasedLinearAttention, | |
| RWKV6Attention, | |
| RWKV7Attention | |
| ) | |
| from fla.models import ( | |
| ABCForCausalLM, | |
| ABCModel, | |
| BitNetForCausalLM, | |
| BitNetModel, | |
| DeltaNetForCausalLM, | |
| DeltaNetModel, | |
| GatedDeltaNetForCausalLM, | |
| GatedDeltaNetModel, | |
| GatedDeltaProductForCausalLM, | |
| GatedDeltaProductModel, | |
| GLAForCausalLM, | |
| GLAModel, | |
| GSAForCausalLM, | |
| GSAModel, | |
| HGRN2ForCausalLM, | |
| HGRN2Model, | |
| HGRNForCausalLM, | |
| LightNetForCausalLM, | |
| LightNetModel, | |
| LinearAttentionForCausalLM, | |
| LinearAttentionModel, | |
| NSAForCausalLM, | |
| NSAModel, | |
| RetNetForCausalLM, | |
| RetNetModel, | |
| RWKV6ForCausalLM, | |
| RWKV6Model, | |
| RWKV7ForCausalLM, | |
| RWKV7Model, | |
| TransformerForCausalLM, | |
| TransformerModel | |
| ) | |
| __all__ = [ | |
| 'ABCAttention', | |
| 'Attention', | |
| 'BasedLinearAttention', | |
| 'BitAttention', | |
| 'DeltaNet', | |
| 'GatedDeltaNet', | |
| 'GatedDeltaProduct', | |
| 'GatedLinearAttention', | |
| 'GatedSlotAttention', | |
| 'HGRNAttention', | |
| 'HGRN2Attention', | |
| 'LightNetAttention', | |
| 'LinearAttention', | |
| 'MultiScaleRetention', | |
| 'NativeSparseAttention', | |
| 'ReBasedLinearAttention', | |
| 'RWKV6Attention', | |
| 'RWKV7Attention', | |
| 'ABCForCausalLM', | |
| 'ABCModel', | |
| 'BitNetForCausalLM', | |
| 'BitNetModel', | |
| 'DeltaNetForCausalLM', | |
| 'DeltaNetModel', | |
| 'GatedDeltaNetForCausalLM', | |
| 'GatedDeltaNetModel', | |
| 'GatedDeltaProductForCausalLM', | |
| 'GatedDeltaProductModel', | |
| 'GLAForCausalLM', | |
| 'GLAModel', | |
| 'GSAForCausalLM', | |
| 'GSAModel', | |
| 'HGRNForCausalLM', | |
| 'HGRNModel', | |
| 'HGRN2ForCausalLM', | |
| 'HGRN2Model', | |
| 'LightNetForCausalLM', | |
| 'LightNetModel', | |
| 'LinearAttentionForCausalLM', | |
| 'LinearAttentionModel', | |
| 'NSAForCausalLM', | |
| 'NSAModel', | |
| 'RetNetForCausalLM', | |
| 'RetNetModel', | |
| 'RWKV6ForCausalLM', | |
| 'RWKV6Model', | |
| 'RWKV7ForCausalLM', | |
| 'RWKV7Model', | |
| 'TransformerForCausalLM', | |
| 'TransformerModel', | |
| ] | |
| __version__ = '0.1.2' | |