܄

MiniMax will release the first large model of MoE in China

【数据猿导读】 MiniMax will release the first large model of MoE in China

MiniMax will release the first large model of MoE in China

On December 28, Wei Wei, vice president of MiniMax, a large model startup company in China, revealed at a sub-forum of the Digital China Forum and Digital Development Forum that the first large model based on MoE(Mixture of Experts) architecture will be released in the near future, marking OpenAI GPT-4. MoE, which stands for Expert Blending, is a deep learning technique that combines multiple models directly together to speed up model training and achieve better predictive performance. A recent paper published by researchers from Google, UC Berkeley, MIT, and other institutions demonstrates that the combination of MoE and instruction tuning can significantly improve the performance of large language models.


来源:DIYuan

声明:数据猿尊重媒体行业规范,相关内容都会注明来源与作者;转载我们原创内容时,也请务必注明“来源:数据猿”与作者名称,否则将会受到数据猿追责。

刷新相关文章

Lenovo launches Personal Agent, opens local large model API
Lenovo launches Personal Agent, opens local large ...
Quark App launches Health large model application
Quark App launches Health large model application ...
Nio released the first self-developed intelligent driving chip Shenguet NX9031
Nio released the first self-developed intelligent d...

我要评论

数据猿微信公众号
2023第七届上海AI大会暨医药和医疗创新峰会
2023深圳物联网展
人工智能博览会
FMW2023全球闪存峰值
2023世界农业科技创新大会暨世界农业科技博览会
2024上海世博展览馆
返回顶部