Back to articles

AMD enthusiast builds local LLM workstation achieving 120 tokens/second with Ryzen 9700X and Radeon R9700, seeking optimal model recommendations.

r/LocalLLaMA · April 20, 2026

AI Summary

  • Builder configured high-end local inference rig with AMD Radeon AI PRO R9700 (32GB VRAM) and Ryzen 7 9700X CPU paired with 64GB DDR5 RAM
  • System achieves ~120 tokens/second performance on simple prompts using Qwen3.6-35B-A3B model via LM Studio with Vulkan backend
  • Poster seeks community advice on largest compatible model architectures and whether Q4_K_M quantizations are optimal for their hardware setup

Related Articles

Stay ahead with AI news

Get curated AI news from 200+ sources delivered daily to your inbox. Free to use.

Get Started Free