Search
Results
How LLMs Work, Explained Without Math - miguelgrinberg.com
LLaMA 7B GPU Memory Requirement - Transformers - Hugging Face Forums
[https://discuss.huggingface.co/t/llama-7b-gpu-memory-requirement/34323/6] - - public:mzimmerm
With the optimizers of bitsandbytes (like 8 bit AdamW), you would need 2 bytes per parameter, or 14 GB of GPU memory.