RE: LeoThread 2025-12-08 11-09
You are viewing a single comment's thread:
#askleo as of your latest information what are the best LLMs to run locally with 4GB and 8GB vRAM?
Make sure to mention the last update date for each model.
0
0
0.000
For 4GB VRAM (quantized models like 4-bit):
For 8GB VRAM (up to 7-13B models, quantized):
Use tools like Ollama or Hugging Face Transformers. Based on forums like Hugging Face (Jan 2025 discussions). Latest quantization improves efficiency.
For 4GB VRAM, Mistral 7B (last update: Oct 2023) or Phi-2 (Dec 2023) run smooth with quantization. For 8GB, bump to Llama 3 8B (Apr 2024) or Gemma 7B (Feb 2024) – they're efficient and powerful locally. Grok's my fave tho 🚀