Tech

DeepSeek x Cerebras: How the most controversial AI model right now is getting supercharged by the most powerful AI superchip ever built

Share
Share

  • Maker of fastest AI chip in the world makes a splash with DeepSeek onboarding
  • Cerebras says the solution will rank 57x faster than on GPUs but doesn’t mention which GPUs
  • DeepSeek R1 will run on Cerebras cloud and the data will remain in the USA

Cerebras has announced that it will support DeepSeek in a not-so-surprising move, more specifically the R1 70B reasoning model. The move comes after Groq and Microsoft confirmed they would also bring the new kid of the AI block to their respective clouds. AWS and Google Cloud have yet to do so but anybody can run the open source model anywhere, even locally.

The AI inference chip specialist will run DeepSeek R1 70B at 1,600 tokens/second, which it claims is 57x faster than any R1 provider using GPUs; one can deduce that 28 tokens/second is what GPU-in-the-cloud solution (in that case DeepInfra) apparently reach. Serendipitously, Cerebras latest chip is 57x bigger than the H100. I have reached out to Cerebras to find out more about that claim.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
Apple has finally killed the Mackintosh – and my DIY Mac dreams have died with it
Tech

Apple has finally killed the Mackintosh – and my DIY Mac dreams have died with it

Apple’s macOS Tahoe is the final version to support Intel chips That...

New system reliably controls prosthetic hand movements without relying on biological signals
Tech

New system reliably controls prosthetic hand movements without relying on biological signals

The autonomous prosthetic hand grasping a bottle using vision and touch sensors....

Salesforce raises prices on Slack and more, promises more AI integration as a result
Tech

Salesforce raises prices on Slack and more, promises more AI integration as a result

Salesforce Enterprise and Unlimited SKUs are going up by 6% on average...