Latest news and articles about inference
Total: 1 articles found
Microsoft has launched Maia 200, a TSMC 3nm AI inference chip the company says outperforms Amazon’s Trainium v3 and Google’s TPU v7 on low-precision workloads while improving inference cost-efficiency by about 30% versus its current fleet. The release underscores hyperscalers’ push into custom silicon to reduce reliance on Nvidia GPUs, but success will depend on software tooling, ecosystem adoption and independent benchmarking.