“Back in my day, we overclocked our Pentiums to run Quake faster. Now we’re doing it for language models.”
— Wing, mid-rant, 2022
In the year 2022, a new frontier opened up for hobbyists and developers alike: Large Language Models (LLMs) stopped being pure research tools and started becoming part of the daily build cycle for curious engineers and indie coders. But with that shift came a new dilemma—do you run your AI workloads locally, or pipe them out to the cloud?