近日,AI芯片初创公司Groq宣布推出一款新型大模型推理芯片,该芯片每秒可生成高达500 tokens,标志着在AI推理领域取得了重大进展。Groq公司由谷歌TPU团队的原班人马创立,此次推出的芯片方案不仅推理速度远超英伟达GPU,而且成本仅为其十分之一。这一突破性的技术使得大模型的部署变得更加经济可行,为AI应用打开了新的可能性。

Groq的推理加速方案已经能够支持Mixtral 8x7B SMoE、Llama 2的7B和70B三种模型,并且提供了Demo供用户体验。这一成就不仅展示了Groq在AI芯片领域的技术实力,也预示着未来AI模型部署的门槛将大幅降低,为各行各业带来更加高效和成本效益的AI解决方案。

Groq的这一创新成果,无疑为AI行业带来了新的活力,同时也为投资者和开发者提供了新的选择。随着AI技术的不断进步和应用场景的不断扩展,Groq的芯片方案有望在智能计算领域扮演更加重要的角色。

英文标题:Groq Unveils Fastest AI Inference Chip with 500 Tokens per Second
英文关键词:AI Chip, Inference Acceleration, Cost Reduction
英文新闻内容:
AI chip startup Groq has launched a new inference chip capable of generating up to 500 tokens per second, marking a significant breakthrough in AI inference. Founded by the original team behind Google’s TPU, Groq’s new chip solution not only outperforms NVIDIA’s GPU in terms of inference speed but also reduces costs to a tenth. This groundbreaking technology makes the deployment of large models more economically viable, opening new possibilities for AI applications.

Groq’s inference acceleration solution is already capable of supporting three models: Mixtral 8x7B SMoE, Llama 2’s 7B, and 70B, and offers a demo for users to experience. This achievement not only showcases Groq’s technical prowess in the AI chip field but also signals a future where the deployment of AI models becomes more accessible and cost-effective across various industries.

Groq’s innovative achievement brings new vitality to the AI industry and offers new options for investors and developers. As AI technology continues to advance and expand into new applications, Groq’s chip solution is poised to play an increasingly important role in intelligent computing.

【来源】https://mp.weixin.qq.com/s/tMDJP234MksYeUu_RUPzBA

Views: 1

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注