Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

news pappernews papper
0

Meta公司近期公开了其用于训练最新Llama3大语言模型的24k GPU集群的详细信息。这一消息在人工智能和科技界引起了广泛关注。Meta的AI Research SuperCluster(RSC)在2022年推出,而现在的新集群是RSC的升级版,其规模和性能都有了显著提升。

新发布的24k GPU集群是Meta致力于推动生成式人工智能技术发展的重要举措。该集群包含24,576个Nvidia Tensor Core H100 GPU,较之之前的16,000个Nvidia A100 GPU,这是一个巨大的飞跃。这种升级使得Meta能够支持更大、更复杂的模型,为生成式AI产品的开发开辟了新的可能性。

Meta在博文中详细介绍了新集群的硬件、网络、存储、设计、性能和软件。这些信息不仅展示了Meta在人工智能基础设施方面的投入和实力,也为其他研究机构和公司提供了宝贵的参考。

随着Llama3模型的训练,Meta有望在自然语言处理、语音识别和图像生成等领域取得突破性进展。这一系列的创新和进步,标志着人工智能技术正朝着更高效、更智能的方向发展。

英文标题:Meta Unveils Detailed Info on 24k GPU Clusters for Llama3 Training

英文关键词:GPU cluster, Meta AI, Llama3 model

英文新闻内容:
Meta has recently unveiled detailed information about its 24k GPU cluster, which will be used to train the latest Llama3 large language model. This announcement has sparked significant interest in the AI and tech communities. The upgraded 24k GPU cluster, part of Meta’s AI Research SuperCluster (RSC) launched in 2022, represents a significant leap in scale and performance.

The new cluster, with its 24,576 Nvidia Tensor Core H100 GPUs, marks a significant increase from the previous 16,000 Nvidia A100 GPUs. This upgrade enables Meta to support larger and more complex models, paving the way for advancements in generative AI product development.

Meta’s blog post provides a detailed overview of the hardware, networking, storage, design, performance, and software of the new cluster. This information not only showcases Meta’s investment and capabilities in AI infrastructure but also serves as a valuable reference for other research institutions and companies.

With the training of the Llama3 model underway, Meta is poised to make breakthroughs in areas such as natural language processing, speech recognition, and image generation. This series of innovations and advancements signifies the ongoing development of AI technology towards more efficient and intelligent applications.

【来源】https://www.datacenterdynamics.com/en/news/meta-reveals-details-of-two-new-24k-gpu-ai-clusters/

Views: 1

0

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注