ShallowFF
PublicZeta implemantion of "Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers"
artificial-intelligenceattentionattention-is-all-you-needattention-mechanismattention-mechanismsfeedforwardtransformertransformer-encodertransformer-modelstransformers-models
Creat:2023-11-20T11:49:13
Update:2025-01-27T11:00:18
https://discord.gg/Yx5y5VBahs
12
Stars
0
Stars Increase