23.2 C
New York
Wednesday, August 21, 2024

Microsoft’s new Phi 3.5 LLM fashions surpass Meta and Google



Throughout benchmarks ranking fashions on reasoning and multilingual expertise, akin to BigBench, MMLU, and ARC Problem, the MoE-instruct mannequin, though with fewer parameters than rivals (6.6 billion) carried out higher than Llama 3.1-8B-instruct, Gemma 2-9b-It, and Gemini 1.5-Flash. Nonetheless, it couldn’t match the efficiency of OpenAI’s GPT-4o-mini-2024-07-18 (chat).

Nonetheless, the corporate identified that the mannequin continues to be basically restricted by its measurement for sure duties.

“The mannequin merely doesn’t have the capability to retailer an excessive amount of factual information, due to this fact, customers might expertise factual incorrectness,” it stated, including that this weak spot will be resolved by augmenting Phi-3.5 with a search engine, significantly when utilizing the mannequin beneath RAG settings.



Supply hyperlink

Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles