Currently, the technical achievement paper "Every FLOP Counts: Scaling 300 Billion Parameter Mixed Expert LING Big Models without Advanced GPUs" from Ant Group's Ling team has been published on the pre-print Arxiv platform. The paper shows that Ant Group has launched two different sizes of MoE big language models - Ling-Lite and Ling-Plus, with the former having a parameter scale of 16.8 billion (2.75 billion activation parameters) and the Plus base model having a parameter scale of up to 290 billion ( 28.8 billion activation parameters), both performance to reach the industry leading level. (Perking synthesized)
❯
Search
Scan to open current page
Top
Checking in, please wait
Click for today's check-in bonus!
You have earned {{mission.data.mission.credit}} points today!
My Coupons
-
¥CouponsLimitation of useExpired and UnavailableLimitation of use
before
Limitation of usePermanently validCoupon ID:×Available for the following products: Available for the following products categories: Unrestricted use:Available for all products and product types
No coupons available!
Unverify
Daily tasks completed:
