NVIDIA RTX 4090 cannot run Llama 3 70B (70.00B) in this configuration. The model requires 35.0GB but only 24.0GB is available, leaving you 11.0GB short.
Consider using a more aggressive quantization (Q4_K_M, Q3_K_M) to reduce VRAM requirements, or upgrade to a GPU with more VRAM. Cloud GPU services like RunPod or Vast.ai offer affordable options.