Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Launchpad
Be early to the next big token project
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Google's first native multimodal embedding model, Gemini Embedding 2, released: enabling machines to "understand" information
IT Home, March 11 — Google announced the release of the new Gemini Embedding2 model early this morning, Beijing time. This is Google’s first native multimodal embedding model, capable of mapping text, images, videos, and documents into the same embedding space.
Embedding models differ from generative models. While models like Gemini3 are mainly used for content generation, embedding models are used for understanding data. They convert text, images, or videos into mathematical forms like vectors, making it easier for machines to read and analyze.
Through semantic search, classification, and clustering, these models can understand semantic relationships, often providing more accurate and context-aware information than traditional keyword searches.
According to IT Home, Google’s earliest embedding models only supported text. Gemini Embedding2 now supports text, images, videos, audio, and documents, and can recognize semantic intent in 100 languages.
Processing limitations for different data types are as follows:
Google stated in a blog that the new model simplifies complex data processing workflows and enhances multimodal application capabilities. Use cases include retrieval-augmented generation (RAG), semantic search, sentiment analysis, and data clustering.
The model can also accept multiple input types simultaneously in a single request, such as “image + text,” to analyze relationships between different media types.
For example, during litigation evidence collection, Gemini embedding models can help legal professionals quickly find key evidence. Testing shows that multimodal embeddings improve retrieval accuracy and recall across millions of records, as well as enhance image and video search results.
Gemini Embeddings2 (gemini-embedding-2-preview) is now available for public preview via the Gemini API and Vertex AI. Meanwhile, gemini-embedding-001 remains available for applications that only process text.