←back to thread

7 points cgel | 1 comments | | HN request time: 0.209s | source
1. cgel ◴[] No.45755526[source]
We have trained a completely attention-free LLM whose performance is competitive with state-of-the-art models. This model, which we call Brumby-14B-Base, has a familiar Transformer-style architecture, except it uses power retention layers instead of attention layers. It is available on Huggingface.