Deepseek x brains: how the most controversial model is at this time being supercharged by the most powerful superchip ever built



  • The fastest fastest manufacturer in the world makes a splash with a brief welcome on board
  • Brains
  • Deepseek R1 will be executed in Cloud brains and the data will remain in the US.

Brains has announced that it will support Depseek in a not -so -surprising movement, more specifically the R1 70b reasoning model. The movement occurs after Groq and Microsoft confirmed that they would also take the new boy from the AI ​​block to their respective clouds. AWS and Google Cloud have not yet done so, but anyone can run the open source model anywhere, even locally.

The inference chip specialist will execute Deepseek R1 70b at 1,600 tokens/second, which states that it is 57 times faster than any R1 provider that uses GPU; One can deduce that 28 tokens/second is what the GPU solution in the cloud (in that case deepfra) apparently reaches. By chance, the last brain chip is 57 times larger than H100. I have communicated with the brains for more information about that statement.

Leave a Comment

Your email address will not be published. Required fields are marked *