I recently made a little tool for people interested in running local LLMs to figure out if their hardware is able to run an LLM in GPU memory.
replies(10):
Feature request - Have a leaderboard of LLM for x/y/z tasks or pull it from existing repo. Suggest the best model for given GPU for x/y/z task.
If there is better model which my GPU can run, why should I go for the lowest?