AI Startup Groq Debuts in the Cloud

Groq has become the second start-up to debut its deep learning chip in the cloud, EETimes reported. Its tensor streaming processor (TSP) for AI inference is now available via cloud service provider Nimbix for “selected customers”.

Late last year, Graphcore became the first startup to see its AI chip become available in the cloud through Microsoft’s Azure. Nimbix is now offering Groq’s tensor streaming processor.

TOPICS
  • bit_user
    As seems to be the trend, it's heavily-dependent on on-chip memory. The pic of the PCIe card doesn't show any off-chip memory, though perhaps there's some HBM2 under the IHS? Otherwise, you might hit a (performance) wall, when your model tries to scale beyond what they can fit on-chip.

    I would also worry about the energy usage resulting from all of the on-chip data movement, since the on-chip memory is supposedly organized in a global pool.

    The top-line numbers are impressive, though.
    Reply