Even my 4-year-old M1 Pro can run a quantized Deepseek R1 pretty well. Sure, full-scale productizing these models is hard work (and the average "just-make-shovels" startups are failing hard at this), but we'll 100% get there in the next 1-2 years.
Those small models suck. You need the big guns to get those "amazing" coding agents.
Local for emotional therapy. Big guns to generate code. Local to edit generated code once it is degooped and worth something.