What is inference? Explaining the massive new shift in AI computing
- 2 hours ago
- 1 min read

Global capital spending on inference infrastructure—including chips, data centers and networking hardware—is expected to surpass capital expenditures on training for the first time this year, according to research firm Gartner. By 2029, companies will spend nearly twice as much on inferencing as on training ($72 billion vs. $37 billion).
That means big changes in the types of chips tech companies buy. Nvidia became the world’s most valuable company by selling chips, called GPUs, that have the raw processing power needed for model training.
But companies that expect to be doing more inference can achieve performance gains by doing it on chips specialized for that task, said Jacob Feldgoise, a scholar who researches AI at Georgetown University.
Read the full story | WALL STREET JOURNAL


