Is there a reason Llama is getting so much attention compared to say T5 11B?
Not sure how neutral or what benchmarks are used on the following link, but T5 seems to sit a lot higher on this leaderboard?
replies(2):
Not sure how neutral or what benchmarks are used on the following link, but T5 seems to sit a lot higher on this leaderboard?
In the end they are mathematical models, so what would prevent someone from loading T5 into a machine with plenty of RAM (like a server)? Would the codebase truly require that much refactoring? How difficult would it be to rewrite the model arhitecture as a set of mathematical equations (Einstein summation) and reimplement inference for CPU?
Most people don't have the hardware or budget to access these specialized high vram GPUs.