Overview
Google Gemini is Google’s family of multimodal AI models that can process text, images, audio, and video. Nadoo AI supports Gemini through two access paths: Google AI Studio (API key-based, for experimentation and smaller workloads) and Google Vertex AI (service account-based, for production and enterprise use). Key strengths:- Multimodal natively — Process text, images, audio, and video in a single model call
- Large context windows — Gemini Pro supports up to 1M tokens for massive document processing
- Two access paths — Simple API key setup (AI Studio) or enterprise-grade deployment (Vertex AI)
- Embedding support — Dedicated embedding models for knowledge base and RAG
Access Paths
- Google AI Studio
- Google Vertex AI
Best for: Development, experimentation, and smaller production workloads.Google AI Studio provides a simple API key-based access to Gemini models. No Google Cloud project setup is required.
| Feature | Detail |
|---|---|
| Authentication | API Key |
| Pricing | Free tier available; pay-as-you-go |
| Data residency | No regional control |
| SLA | No enterprise SLA |
Setup — Google AI Studio
Get an API Key
Go to aistudio.google.com and click Get API Key. Create a new key or select an existing one.
Configure in Nadoo
Go to Admin > Model Providers > Google Gemini and enter:
| Field | Required | Description |
|---|---|---|
| API Key | Yes | Your Google AI Studio API key |
| Access Path | Yes | Select AI Studio |
Setup — Google Vertex AI
Create a Google Cloud Project
Set up a project in the Google Cloud Console and enable the Vertex AI API.
Create a Service Account
Create a service account with the Vertex AI User role (
roles/aiplatform.user). Download the JSON key file.Configure in Nadoo
Go to Admin > Model Providers > Google Gemini and enter:
| Field | Required | Description |
|---|---|---|
| Access Path | Yes | Select Vertex AI |
| Project ID | Yes | Your Google Cloud project ID |
| Region | Yes | The Google Cloud region (e.g., us-central1) |
| Service Account JSON | Yes | Upload or paste the service account key JSON |
Available Models
Chat / LLM
| Model | Context Window | Best For |
|---|---|---|
gemini-2.0-flash | 1M tokens | Fast, efficient multimodal tasks |
gemini-2.0-pro | 1M tokens | Complex reasoning with massive context |
gemini-1.5-pro | 1M tokens | Long document analysis, multimodal processing |
gemini-1.5-flash | 1M tokens | Cost-efficient, fast responses |
Embedding
| Model | Dimensions | Best For |
|---|---|---|
text-embedding-004 | 768 | General-purpose embeddings for RAG |
text-multilingual-embedding-002 | 768 | Multilingual embeddings |
Capabilities
Chat Completion
Conversational AI with streaming and function calling support.
Vision
Analyze images alongside text — charts, documents, screenshots, and photos.
Embeddings
Generate vector embeddings for semantic search and knowledge base indexing.
Long Context
Process up to 1 million tokens — entire codebases, books, or document collections in one request.
Multimodal
Process text, images, audio, and video in a single model call.
Function Calling
Invoke tools with structured arguments for agentic workflows.
When to Use Google Gemini
Massive context processing
Massive context processing
Gemini’s 1M token context window is the largest available. Use it when you need to process entire books, large codebases, or extensive document collections without chunking.
Multimodal tasks
Multimodal tasks
When your application needs to analyze images, process audio, or understand video alongside text, Gemini’s native multimodal support avoids the need for separate processing pipelines.
Google Cloud ecosystem
Google Cloud ecosystem
If your infrastructure is on Google Cloud, Vertex AI integrates with IAM, VPC, Cloud Logging, and other GCP services for a seamless enterprise experience.
Cost-efficient development
Cost-efficient development
Google AI Studio offers a generous free tier for experimentation and development, making it easy to prototype before committing to production costs.
Recommended Models by Use Case
| Use Case | Recommended Model | Reason |
|---|---|---|
| General chatbot | gemini-2.0-flash | Fast, multimodal, cost-efficient |
| Complex analysis | gemini-2.0-pro | Highest reasoning capability |
| Long document Q&A | gemini-1.5-pro | 1M token context for massive documents |
| High-volume workloads | gemini-1.5-flash | Fastest Gemini model |
| Knowledge base search | text-embedding-004 | Good quality embeddings |
| Multilingual RAG | text-multilingual-embedding-002 | Cross-language embedding support |
Environment Variables
When self-hosting, configure Google Gemini via environment variables:AI Studio
Vertex AI
Rate Limits and Pricing
AI Studio
Google AI Studio applies rate limits per API key with a free tier and pay-as-you-go pricing above the free quota.Vertex AI
Vertex AI uses per-project quotas that can be adjusted in the Google Cloud Console. Pricing is per 1,000 characters for input and output.Troubleshooting
403 Permission Denied
403 Permission Denied
AI Studio: Your API key is invalid or revoked. Generate a new key at aistudio.google.com.Vertex AI: Your service account lacks the Vertex AI User role. Check IAM permissions in the Google Cloud Console.
429 Resource Exhausted
429 Resource Exhausted
You have exceeded rate limits. Wait and retry, or request a quota increase in the Google Cloud Console (Vertex AI only).
Model not available
Model not available
The model may not be available in your selected region (Vertex AI) or may not be enabled for your API key (AI Studio). Check model availability documentation.