Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

在上海浦东滨江公园观赏外滩建筑群-20240824在上海浦东滨江公园观赏外滩建筑群-20240824
0

Microsoft Unveils Phi-3.5: A New Generation of AI Models withMultimodal Capabilities

Seattle, WA – Microsoft has announced the release ofPhi-3.5, a new generation of AI models designed to push the boundaries of language understanding and generation. The Phi-3.5 series comprises threedistinct models: Phi-3.5-mini-instruct, Phi-3.5-MoE-instruct, and Phi-3.5-vision-instruct, each tailored for specific tasks and capabilities.

Phi-3.5-mini-instruct: This model, with approximately 38.2 billion parameters, is optimized for fast inference tasks. Designedto follow instructions, it excels in code generation, solving mathematical problems, and logical reasoning. Its ability to handle 128k token context length makes it suitable for processing long text data. In benchmark tests, Phi-3.5-mini-instruct outperformed models of similar size, including Llama-3.1-8B-instruct and Mistral-7B-instruct, in tasks like long-context code understanding.

Phi-3.5-MoE-instruct: This model, boasting 41.9 billion parameters, employs a Mixture-of-Experts (MoE) architecture, combining multiple specialized models for different tasks. This allows it to handle complex multi-language and multi-task scenarios. Phi-3.5-MoE-instruct excels in code, mathematics, and multi-language understanding, often outperforming larger models in specific benchmarks. It demonstrates remarkable performance in the RepoQA benchmark and surpasses GPT-40 mini in the 5-shot MMLU (Massive Multitask Language Understanding) benchmark across various disciplines.

Phi-3.5-vision-instruct: This model, with41.5 billion parameters, integrates text and image processing capabilities, enabling it to handle multimodal data. It is particularly adept at general image understanding, Optical Character Recognition (OCR), chart and table comprehension, and video summarization. With 128k token context length support, Phi-3.5-vision-instruct can manage complex multi-frame visual tasks. The model is trained on a combination of synthetic and curated public datasets, emphasizing high-quality, reasoning-intensive data.

Open Source and Performance: All Phi-3.5 models are released under the MIT open-source license, allowing researchers anddevelopers to access and utilize them freely. The models have demonstrated impressive performance across various benchmarks, surpassing existing models like GPT-40, Llama 3.1, and Gemini Flash in key areas.

Significance and Impact: The release of Phi-3.5 marks a significant advancement in the field of AI,particularly in the development of large language models. Its multimodal capabilities, combined with its impressive performance and open-source nature, have the potential to revolutionize various industries, including research, education, healthcare, and entertainment.

Future Directions: Microsoft is actively working on further enhancing the capabilities of Phi-3.5 models. The company is exploring ways to improve their performance, expand their functionalities, and make them more accessible to a wider audience.

Conclusion: Phi-3.5 represents a significant step forward in the development of AI models, offering a powerful and versatile tool for various applications. Its open-source nature fosters collaboration andinnovation within the AI community, paving the way for exciting advancements in the field. As Microsoft continues to refine and expand the capabilities of Phi-3.5, we can expect to see even more groundbreaking applications emerge in the future.

【source】https://ai-bot.cn/phi-3-5/

Views: 0

0

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注