NVIDIA RTX 3090 Ti cannot run Qwen 2.5 14B (14.00B) in this configuration. The model requires 28.0GB but only 24.0GB is available, leaving you 4.0GB short.
Consider using a more aggressive quantization (Q4_K_M, Q3_K_M) to reduce VRAM requirements, or upgrade to a GPU with more VRAM. Cloud GPU services like RunPod or Vast.ai offer affordable options.