Does that also mean 6GB VRAM?
And does that include Alpaca models like this? https://huggingface.co/elinas/alpaca-30b-lora-int4
replies(2):
And does that include Alpaca models like this? https://huggingface.co/elinas/alpaca-30b-lora-int4
If you want to run larger Alpaca models on a low VRAM GPU, try FlexGen. I think https://github.com/oobabooga/text-generation-webui/ is one of the easier ways to get that going.