Are all these data centers really going to be running at full capacity with open models like Qwen 3.6 27B that have performance approaching frontier, but can run on consumer hardware? Sure, it’s slow as of now, though there are tweaks to optimize it, and how long until we see open models that run reasonably fast and give frontier models a run for their money? My company MacBook can run models like this, so will there be a point where companies stop paying hundreds per user per month for cloud AI and have devs run open models on the laptops they already have? I definitely won’t be surprised if that’s the case.
Are all these data centers really going to be running at full capacity with open models like Qwen 3.6 27B that have performance approaching frontier, but can run on consumer hardware? Sure, it’s slow as of now, though there are tweaks to optimize it, and how long until we see open models that run reasonably fast and give frontier models a run for their money? My company MacBook can run models like this, so will there be a point where companies stop paying hundreds per user per month for cloud AI and have devs run open models on the laptops they already have? I definitely won’t be surprised if that’s the case.