Developer Guide

Gemini Omni API

The Gemini Omni API isn't public yet. Here's what we know about how it will work, what to expect, and how to prepare.

Gemini Omni API Status

As of May 2026, the Gemini Omni API is not yet publicly available. Google is expected to announce Gemini Omni at Google I/O 2026 (May 19), but API access typically follows the announcement by weeks or months.

This page will be updated with: - Official API endpoints and documentation links - Authentication setup (API keys, OAuth) - SDK availability (Python, Node.js, REST) - Pricing details - Rate limits and quotas

Expected API Capabilities

Based on industry reports about Gemini Omni, here is what the API will likely support:

Video Generation Generate videos from text prompts or images. Expected endpoints similar to existing Google AI APIs — POST a request with a prompt and optional image, receive a video URL or base64 output.

Chat-Based Editing Conversational editing sessions where you iteratively modify generated videos through natural language. Most current APIs (Runway, Kling) use a single prompt → single output model, so this would be a different approach.

Multimodal Input Combine text, images, and possibly audio in a single API call. The API would accept a multipart request with different content types.

Object Manipulation Select and replace specific objects within generated video frames. This would require the API to support region-based operations.

How to Access the Gemini Omni API (When Available)

1

Get a Google Cloud Account

You will need a Google Cloud project with billing enabled. Gemini APIs typically require Google AI Studio or Vertex AI access.

2

Enable the API

Navigate to Google AI Studio or the Cloud Console and enable the Gemini Omni API for your project.

3

Get API Credentials

Generate an API key (for Google AI Studio) or set up a service account (for Vertex AI production use).

4

Install the SDK

Google will likely provide official SDKs for Python (@google/generative-ai) and Node.js (@google-cloud/vertexai).

5

Make Your First Request

Generate a test video using the code examples below. We'll update these with the official API once it ships.

Expected Code Example (Preview)

Based on Google's existing Gemini API patterns, here is what the Omni video generation API might look like:

Python: import google.generativeai as genai

genai.configure(api_key="YOUR_API_KEY")

model = genai.GenerativeModel("gemini-omni")

response = model.generate_content(
    "A golden retriever running on a beach at sunset, cinematic 4K",
    generation_config={
        "modality": "video",
        "duration_seconds": 5,
        "resolution": "1080p",
    }
)

# response contains a video file URI or bytes video_url = response.candidates[0].content.parts[0].file_uri

Node.js: import { GoogleGenerativeAI } from "@google/generative-ai";

const genAI = new GoogleGenerativeAI("YOUR_API_KEY");
const model = genAI.getGenerativeModel({ model: "gemini-omni" });
const result = await model.generateContent({
  contents: [{
    parts: [{
      text: "A golden retriever running on a beach at sunset, cinematic 4K"
    }]
  }],
  generationConfig: {
    modality: "video",
    durationSeconds: 5,
    resolution: "1080p",
  }
});

Note: This is speculative based on current Gemini API patterns. The actual API may differ significantly.

Pricing (Expected)

Google has not announced pricing for the Gemini Omni API. Based on pricing for existing Google AI models and competitor pricing:

• Video generation APIs typically charge per second of generated video • Current market rate: $0.02–$0.10 per second of video • Google may offer a free tier with limited monthly generation minutes • Enterprise (Vertex AI) pricing is usually higher than consumer (AI Studio) pricing

We will update this section with exact pricing once Google announces it.

Alternatives (Available Now)

While you wait for the Gemini Omni API, these video generation APIs are available today:

• Kling AI API — Strong physics-based motion, available through Kuaishou's developer platform • Runway Gen-4 API — Popular for cinematic video, available at runwayml.com • Veo API — Google's current video model, available through Google AI Studio

Explore More

Frequently Asked Questions

Is the Gemini Omni API available?
Not yet. It is expected to be released after Google I/O 2026 (May 19). No public API access exists as of May 2026.
How much will the Gemini Omni API cost?
Pricing has not been announced. Based on competitor pricing, expect per-second charges in the $0.02–$0.10 range. Google may offer a free tier.
Will there be a Python SDK for Gemini Omni?
Almost certainly. Google provides Python and Node.js SDKs for all their AI models. The Gemini Omni API will likely extend the existing google-generativeai Python package.
Can I use Gemini Omni on Vertex AI?
We expect Gemini Omni to be available on both Google AI Studio (for prototyping) and Vertex AI (for production deployments). This is consistent with how Google releases other Gemini models.
What video formats will the API support?
Based on current Google video API patterns, expect MP4 output with configurable resolution (720p, 1080p) and duration options. Exact specs will be confirmed at launch.

Ready to Generate AI Videos?

Try our AI video generator today. Generate videos from text or images in your browser.

Start Generating →