NVIDIA H100 SXM cannot run Llama 3.1 405B (405.00B) in this configuration. The model requires 810.0GB but only 80.0GB is available, leaving you 730.0GB short.
Consider using a more aggressive quantization (Q4_K_M, Q3_K_M) to reduce VRAM requirements, or upgrade to a GPU with more VRAM. Cloud GPU services like RunPod or Vast.ai offer affordable options.