Without a doubt there is a strong momentum for AI inferencing and therefore for new AI inferencing hardware. We are beyond the point where AI inferencing has been dominated by Nvidia, AMD, Intel and maybe Groq. We see new vendors like MiTac, Nebius and Positron who are focusing purely on AI inferencing which allows them to focus on AI inferencing features. They are not trying to beat Nvidia in their core business - AI Training but they rather focus on AI inferencing market which has big potential.

I was able to meet with Positron only briefly and few interesting facts about them:

  • Co-founded by ex-Groq and ex-LambdaLabs, 30 employees
  • They just went through series A funding
  • Delivering their own AI inferencing product to traditional datacenters
  • They are not focusing on speed - they focus on performance/W and performance/$
  • They see future of inferencing on agentic MoE architecture - collection of small LLMs
  • At this stage they are focusing on larger deployments
    • Offering AI server with 8x Positron Archer Transformer Accelerators, each with 32GB HBM

I will have a follow up with Positron focusing on their technical architecture in following few weeks so I will follow up with more detailed post. Very exciting to see new hardware vendors - this is exciting field opened to new companies.