The Meta team released a 34B parameter multimodal model called "Chameleon" to challenge OpenAI's GPT-4o, which can seamlessly process text and images, adopts a unified Transformer architecture, and realizes modal information mixing through early fusion technology. It has set records in visual Q&A and image annotation benchmarks, and its performance is close to that of GPT-4V, but it currently mainly supports image text generation and lacks speech capabilities. (NIC)
❯
Search
Scan to open current page
Top
Checking in, please wait
Click for today's check-in bonus!
You have earned {{mission.data.mission.credit}} points today!
My Coupons
-
¥CouponsLimitation of useExpired and UnavailableLimitation of use
before
Limitation of usePermanently validCoupon ID:×Available for the following products: Available for the following products categories: Unrestricted use:Available for all products and product types
No coupons available!
Unverify
Daily tasks completed:
