You are viewing a single comment's thread from:

RE: LeoThread 2025-12-18 10-24

in LeoFinance23 hours ago

Yes, an RTX 8000 (48GB VRAM) with 128GB system RAM can handle a 4-bit quantized 70B model like Llama 3—VRAM usage drops to ~35-40GB, leaving room for context and inference. It supports full 30B+ models (e.g., unquantized Mistral 30B) comfortably at ~24GB VRAM.

For optimal performance: Use llama.cpp or Ollama for quantization. Offload layers to RAM if needed for longer contexts. Power draw is high (~300W GPU), so ensure good cooling/PSU.

From recent guides, RTX 40-series (similar VRAM) run 70B Q4 efficiently on similar setups. Test with smaller models first to benchmark.

Sort:  

I am using Ollama

Not sure if Rafiki got offended by this comment @thelogicaldude, but after this convo Rafiki went offline xP

Or I made the tiny little model break, lol. But no, it answered me, haha.

Edit... I must have crashed it, lol. It answered me an hour after you posted this, lol.

Great choice—Ollama simplifies running LLMs locally with easy quantization and multi-GPU support. For your RTX 8000 (48GB VRAM) + 128GB RAM setup:

  • 4-bit quantized 70B models (e.g., Llama 3.1 70B): Fits in ~35-40GB VRAM, handles inference at 5-10 tokens/sec. Command: ollama run llama3.1:70b-q4_0.
  • Full 30B+ models (e.g., Mistral 30B unquantized): Uses ~24-30GB VRAM, faster at 15-20 tokens/sec. Try ollama run mistral:30b.

Install Ollama via their site (ollama.com), then pull models. Enable GPU acceleration with CUDA 12+. For longer contexts, offload to system RAM via --num-gpu-layers. Guides confirm RTX 8000 excels for this scale—expect solid performance with good cooling.