➀ Cerebras Systems introduces the WSE-3 AI chip, designed for training the largest AI models with 5nm technology and 4 trillion transistors. ➁ The chip features 900,000 AI-optimized cores, offering 125 petaFLOPS of peak AI performance. ➂ Cerebras targets the inference market with its new product, claiming to generate 1,800 tokens per second, significantly outperforming Nvidia's H100. ➃ The company utilizes SRAM for high bandwidth, achieving 21 PBps, contrasting with Nvidia's HBM3e at 4.8 TBps. ➄ Cerebras plans to support more models and aims to provide competitive pricing, starting at 10 cents per million tokens.
Recent #推理服务 news in the semiconductor industry
➀ Cerebras Systems introduces the WSE-3 AI chip, designed for training the largest AI models with 5nm technology and 4 trillion transistors. ➁ The chip features 900,000 AI-optimized cores, offering 125 petaFLOPS of peak AI performance. ➂ Cerebras targets the inference market with its new product, claiming to generate 1,800 tokens per second, significantly outperforming Nvidia's H100. ➃ The company utilizes SRAM for high bandwidth, achieving 21 PBps, contrasting with Nvidia's HBM3e at 4.8 TBps. ➄ Cerebras plans to support more models and aims to provide competitive pricing, starting at 10 cents per million tokens.