←back to thread

111 points mirrir | 1 comments | | HN request time: 0.217s | source
Show context
adityashankar ◴[] No.46176854[source]
Due to perverse incentives and the historical nature of models over-claiming accuracy, it's very hard to believe anything until it is open source and can be tested out

that being said, I do very much believe that computational efficiency of models is going to go up [correction] drastically over the coming months, which does pose interesting questions over nvidia's throne

*previously miswrote and said computational efficiency will go down

replies(3): >>46176877 #>>46176899 #>>46177234 #
credit_guy ◴[] No.46176899[source]
Like this?

https://huggingface.co/amd/Zebra-Llama-8B-8MLA-24Mamba-SFT

replies(4): >>46176995 #>>46177440 #>>46177468 #>>46177471 #
1. deepdarkforest ◴[] No.46177468[source]
> which does pose interesting questions over nvidia's throne...

> Zebra-Llama is a family of hybrid large language models (LLMs) proposed by AMD that...

Hmmm