Chat
Ask me anything
Ithy Logo

Detailed History of Google Gemini Models: Context Limits and Pricing

Google's Gemini family of models has undergone significant evolution since its inception, marked by advancements in capabilities, context window sizes, and pricing structures. This comprehensive history details the progression of Gemini models, their context limits, and per-token pricing, drawing from official Google documentation and credible sources.


Early Foundations: LaMDA and PaLM

Before the Gemini series, Google developed foundational language models that paved the way for its current AI offerings:

  • LaMDA (Language Model for Dialogue Applications) - 2021: This model was Google's initial foray into conversational AI, laying the groundwork for future models. While not directly part of the Gemini family, it was a crucial precursor. 1
  • PaLM (Pathways Language Model) - 2022: PaLM represented a significant leap forward, boasting enhanced coding, multilingual, and reasoning skills compared to LaMDA. It was later succeeded by PaLM 2 which was used in Bard. 1

Gemini 1.0 Pro

Gemini 1.0 Pro marked the initial release of the Gemini family, designed for general-purpose AI tasks.

  • Release Date: December 2023. 2
  • Capabilities: It offered text and image reasoning, along with stable and reliable performance for production use, supporting text generation, embeddings, and structured outputs. 3
  • Context Limit: 32,000 tokens. 4
  • Pricing:
    • Free Tier:
      • Input: Free
      • Output: Free
      • Rate Limits: 15 requests per minute (RPM), 32,000 tokens per minute (TPM), 1,500 requests per day (RPD). 5
    • Pay-as-You-Go:
      • Input: $1.25 per 1 million tokens (for prompts <128k tokens). 6
      • Output: $5.00 per 1 million tokens (for prompts <128k tokens). 6
      • Context Caching: $0.3125 per 1 million tokens. 6
      • Prompts longer than 128k tokens:
        • Input: $2.50 per 1 million tokens. 6
        • Output: $10.00 per 1 million tokens. 6
        • Context Caching: $0.625 per 1 million tokens. 6

Gemini 1.5 Series

The Gemini 1.5 series introduced significant advancements, particularly in context window size and multimodal capabilities.

Gemini 1.5 Pro
  • Initial Release Date: May 2024 (announced at Google I/O). 7
  • Updated Version: Gemini-1.5-Pro-002 (Released September 24, 2024). 8
  • Capabilities: Enhanced general performance across text, code, and multimodal tasks. 9
  • Context Limit: 2 million tokens. 10
  • Pricing:
    • Free Tier:
      • Input: Free
      • Output: Free
      • Rate Limits: 2 RPM, 32,000 TPM, 50 RPD. 11
    • Pay-as-You-Go (as of September 2024):
      • Input: $1.25 per 1 million tokens (for prompts <128k tokens). 12
      • Output: $5.00 per 1 million tokens (for prompts <128k tokens). 12
      • Context Caching: $0.3125 per 1 million tokens. 12
      • Prompts longer than 128k tokens:
        • Input: $2.50 per 1 million tokens. 12
        • Output: $10.00 per 1 million tokens. 12
        • Context Caching: $0.625 per 1 million tokens. 12
    • Pricing Update (September 24, 2024):
      • >50% reduction in pricing for prompts <128k tokens. 13
Gemini 1.5 Flash
  • Initial Release Date: May 2024. 14
  • Updated Version: Gemini-1.5-Flash-002 (Released September 24, 2024). 15
  • Capabilities: Optimized for lower latency and faster output. 16
  • Context Limit: 1 million tokens. 17
  • Pricing:
    • Free Tier:
      • Input: Free
      • Output: Free
      • Rate Limits: 15 RPM, 1 million TPM, 1,500 RPD. 18
    • Pay-as-You-Go:
      • Input: $1.25 per 1 million tokens. 19
      • Output: $5.00 per 1 million tokens. 19
      • Context Caching: Free for up to 1 million tokens of storage per hour. 19
    • Rate Limit Update (September 24, 2024):
      • 2x increase in rate limits for Flash models. 20
Gemini 1.5 Flash-8B Experimental (Exp-0924)
  • Release Date: September 24, 2024. 21
  • Capabilities: Experimental updates with improved performance across text and multimodal use cases. 22
  • Context Limit: 1 million tokens. 23
  • Pricing: Same as Gemini 1.5 Flash. 24

Gemini 2.0

  • Announcement Date: December 11, 2024. 25
  • Capabilities: Positioned as a next-generation AI model for the "agentic era," with advanced reasoning and multimodal capabilities. 26
  • Context Limit: Not explicitly mentioned but expected to exceed 2 million tokens based on advancements in the Gemini series. 27
  • Pricing: Details not yet disclosed but expected to align with or improve upon the Gemini 1.5 Pro pricing structure. 28

Key Updates and Milestones

  1. December 2023: Launch of Gemini 1.0 Pro with a 32,000-token context window. 29
  2. May 2024: Introduction of Gemini 1.5 Pro and Flash models at Google I/O, featuring multimodal capabilities and larger context windows. 30
  3. June 2024: Updates to Gemini Code Assist to use the Gemini 1.5 Flash model. 31
  4. September 24, 2024: Release of Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002 with:
    • >50% reduced pricing for 1.5 Pro. 32
    • 2x faster output and 3x lower latency. 33
    • 2x higher rate limits for Flash and ~3x higher for Pro. 34
    • Updated default filter settings. 35
  5. December 11, 2024: Announcement of Gemini 2.0 as the next-generation model. 36

Summary of Pricing Structures

Free Tier:
  • Input and output are free for all models.
  • Rate limits vary by model:
    • Gemini 1.0 Pro: 15 RPM, 32,000 TPM. 37
    • Gemini 1.5 Flash: 15 RPM, 1 million TPM. 38
    • Gemini 1.5 Pro: 2 RPM, 32,000 TPM. 39
Pay-as-You-Go:
  • Input Pricing: $1.25 per 1 million tokens (for prompts <128k tokens). 40
  • Output Pricing: $5.00 per 1 million tokens (for prompts <128k tokens). 40
  • Context Caching: $0.3125 per 1 million tokens. 40
  • Prompts >128k tokens: Higher rates apply. 41

Consumer Access


Notes

  • All pricing and rate limits are subject to updates as per Google's official announcements.
  • Developers can access Gemini models via Google AI Studio or Vertex AI for enterprise use cases.
  • Safety filters are available but not applied by default for the latest Gemini models, allowing developers to customize configurations.
  • Context caching is available for free up to specified limits for models like Gemini 1.5 Flash and Flash-8B.
  • Grounding with Google Search is not supported in any of the models as of the latest updates.
  • Experimental models are periodically released and replaced with newer versions as part of Google's development process.

This detailed history provides a comprehensive overview of the evolution of Google Gemini models, ensuring accuracy and relevance based on official sources. The information is current as of December 11, 2024.

1 LaMDA and PaLM details from Opinion A.

2 Gemini 1.0 Pro release date from Opinion D.

3 Gemini 1.0 Pro capabilities from Opinion C.

4 Gemini 1.0 Pro context limit from Opinion B.

5 Gemini 1.0 Pro free tier rate limits from Opinion B.

6 Gemini 1.0 Pro pay-as-you-go pricing from Opinion B.

7 Gemini 1.5 Pro initial release date from Opinion B.

8 Gemini 1.5 Pro updated version from Opinion B.

9 Gemini 1.5 Pro capabilities from Opinion B.

10 Gemini 1.5 Pro context limit from Opinion B.

11 Gemini 1.5 Pro free tier rate limits from Opinion C.

12 Gemini 1.5 Pro pay-as-you-go pricing from Opinion B.

13 Gemini 1.5 Pro pricing update from Opinion B.

14 Gemini 1.5 Flash initial release date from Opinion B.

15 Gemini 1.5 Flash updated version from Opinion B.

16 Gemini 1.5 Flash capabilities from Opinion B.

17 Gemini 1.5 Flash context limit from Opinion B.

18 Gemini 1.5 Flash free tier rate limits from Opinion C.

19 Gemini 1.5 Flash pay-as-you-go pricing from Opinion B.

20 Gemini 1.5 Flash rate limit update from Opinion B.

21 Gemini 1.5 Flash-8B Experimental release date from Opinion B.

22 Gemini 1.5 Flash-8B Experimental capabilities from Opinion B.

23 Gemini 1.5 Flash-8B Experimental context limit from Opinion B.

24 Gemini 1.5 Flash-8B Experimental pricing from Opinion B.

25 Gemini 2.0 announcement date from Opinion B.

26 Gemini 2.0 capabilities from Opinion B.

27 Gemini 2.0 context limit expectation from Opinion B.

28 Gemini 2.0 pricing expectation from Opinion B.

29 Gemini 1.0 Pro launch date from Opinion B.

30 Gemini 1.5 Pro and Flash introduction date from Opinion B.

31 Gemini Code Assist update from Opinion B.

32 Gemini 1.5 Pro pricing reduction from Opinion B.

33 Gemini 1.5 Pro and Flash output speed and latency improvements from Opinion B.

34 Gemini 1.5 Pro and Flash rate limit increases from Opinion B.

35 Gemini 1.5 Pro and Flash filter settings update from Opinion B.

36 Gemini 2.0 announcement date from Opinion B.

37 Gemini 1.0 Pro free tier rate limits from Opinion B.

38 Gemini 1.5 Flash free tier rate limits from Opinion B.

39 Gemini 1.5 Pro free tier rate limits from Opinion C.

40 Gemini pay-as-you-go pricing from Opinion B.

41 Gemini pay-as-you-go pricing for prompts >128k tokens from Opinion B.

42 Gemini Advanced pricing from Opinion A.

43 Gemini Advanced access to Gemini 1.0 Ultra from Opinion A.

44 Gemini Advanced included features from Opinion A.


December 16, 2024
Ask Ithy AI
Download Article
Delete Article