AI芯片初创公司Groq,由谷歌TPU团队的核心成员创立,近期推出了一款革命性的推理芯片,能够在每秒内生成高达500个tokens,标志着AI计算速度的显著提升。这款自研芯片旨在加速AI模型的推理过程,据公司介绍,其性能表现比英伟达的GPU提升了10倍,同时成本降低了90%,为大模型的广泛部署提供了可能。

Groq的创新解决方案使得即使是计算需求庞大的Mixtral 8x7B SMoE、Llama 2的7B和70B这三种模型,也能轻松应对。这一突破性的技术进展使得开发者和企业能够更加高效、经济地运行大规模的AI模型,为人工智能的应用打开了新的大门。

目前,Groq已提供Demo体验,让潜在用户和开发者有机会直接感受这款芯片的卓越性能。这一创新成果无疑将对AI行业的未来发展产生深远影响,为快速、经济的模型推理设定新的行业标准。来源:量子位。

英语如下:

Title: “Groq’s Disruptive AI Chip: 500 Tokens/Second, 10x Faster Than NVIDIA, 1/10th the Cost”

Keywords: Groq, AI chip, inference speed

News Content:

Groq, an AI chip startup founded by key members of Google’s TPU team, has recently unveiled a groundbreaking inference chip that can process up to 500 tokens per second, signifying a significant boost in AI computing speed. The proprietary chip is designed to accelerate the inference process of AI models, with the company claiming it outperforms NVIDIA’s GPUs by a factor of 10 and reduces costs by 90%, paving the way for widespread deployment of large models.

Groq’s innovative solution enables it to handle compute-intensive models such as Mixtral 8x7B SMoE, Llama 2’s 7B, and 70B with ease. This breakthrough advancement allows developers and businesses to run large-scale AI models more efficiently and cost-effectively, opening new doors for AI applications.

Currently, Groq is offering demos to potential users and developers, giving them a firsthand experience of the chip’s exceptional performance. This innovative achievement is set to have a profound impact on the future of the AI industry, establishing new benchmarks for fast and cost-effective model inference. Source: Quantum Bit.

【来源】https://mp.weixin.qq.com/s/tMDJP234MksYeUu_RUPzBA

Views: 1

发表回复

您的电子邮箱地址不会被公开。 必填项已用 * 标注