Google Rolls Out Gemini 1.5 Featuring Experimental 1M Token Context
Google's launch of Gemini 1.5 represents a significant leap in AI context window sizes, with an 'experimental' capability to process up to a million tokens. This feature extends the understanding of exceedingly long text passages, potentially outperforming predecessors like Claude 2.1 and GPT-4 Turbo.
Gemini 1.5's robustness in long-context retrieval tasks is attributed to a novel Mixture-of-Experts (MoE) architecture, which leverages a network of specialized 'expert' neural networks. This results in improved efficiency and response quality, as demonstrated by Google's examples involving extensive text and film summaries.
While currently available to developers and businesses in a limited preview, Gemini 1.5's public release is slated to include a 128,000 token context window. If successful, this model could redefine AI's capacity for complex text interpretation.