Build interactive applications using Google Gemini's multi-turn conversation and extended interaction capabilities. For creating conversational AI experiences.
npx claudepluginhub joshuarweaver/cascade-ai-ml-engineering --plugin google-gemini-gemini-skills-1Use this skill when building applications with Gemini API hosted models, including Gemini and Gemma 4, working with multimodal content (text, images, audio, video), implementing function calling, using structured outputs, or needing current model specifications. Covers SDK usage (google-genai for Python, @google/genai for JavaScript/TypeScript, com.google.genai:google-genai for Java, google.golang.org/genai for Go), model selection, and API capabilities.
Use this skill when writing code that calls the Gemini API for text generation, multi-turn chat, multimodal understanding, image generation, streaming responses, background research tasks, function calling, structured output, or migrating from the old generateContent API. This skill covers the Interactions API, the recommended way to use Gemini models and agents in Python and TypeScript.
Use this skill when building real-time, bidirectional streaming applications with the Gemini Live API. Covers WebSocket-based audio/video/text streaming, voice activity detection (VAD), native audio features, function calling, session management, ephemeral tokens for client-side auth, and all Live API configuration options. SDKs covered - google-genai (Python), @google/genai (JavaScript/TypeScript).
Guides the usage of Gemini API on Google Cloud Vertex AI with the Gen AI SDK. Use when the user asks about using Gemini in an enterprise environment or explicitly mentions Vertex AI. Covers SDK usage (Python, JS/TS, Go, Java, C#), capabilities like Live API, tools, multimedia generation, caching, and batch prediction.
A library of skills for the Gemini API, SDK and model interactions.
LLMs have fixed knowledge, being trained at a specific point in time. Software dev is fast paced and changes often, where new libraries are launched every day and best practices evolve quickly.
This leaves a knowledge gap that language models can't solve on their own. For example, models don't know about themselves when they're trained, and they aren't necessarily aware of subtle changes in best practices (like thought circulation) or SDK changes.
Skills are a lightweight technique for adding relevant context to your agents. This repo contains skills related to building apps powered by the Gemini API.
Our evaluations found that adding this skill improved an agent's ability to generate correct API code following best practices to 87% with Gemini 3 Flash and 96% with Gemini 3 Pro.
| Skill | Description |
|---|---|
gemini-api-dev | Skill for developing Gemini-powered apps. Provides the best practices for building apps that use the Gemini API. |
vertex-ai-api-dev | Skill for developing Gemini-powered apps on Google Cloud Vertex AI using the Gen AI SDK. Covers tools, multimodal generation, caching, and batch prediction. |
gemini-live-api-dev | Skill for building real-time, bidirectional streaming apps with the Gemini Live API. Covers WebSocket-based audio/video/text streaming, voice activity detection, native audio features, function calling, and session management. |
gemini-interactions-api | Skill for building apps with the Gemini Interactions API. Covers text generation, multi-turn chat, streaming, function calling, structured output, image generation, Deep Research agents, deprecated model guardrails, and both Python and TypeScript SDKs. |
You can browse and install skills using either the Vercel skills CLI or the Context7 skills CLI.
# Interactively browse and install skills.
npx skills add google-gemini/gemini-skills --list
# Install a specific skill (e.g., gemini-api-dev).
npx skills add google-gemini/gemini-skills --skill gemini-api-dev --global
# Interactively browse and install skills.
npx ctx7 skills install /google-gemini/gemini-skills
# Install a specific skill (e.g., vertex-ai-api-dev).
npx ctx7 skills install /google-gemini/gemini-skills vertex-ai-api-dev
This is not an officially supported Google product. This project is not eligible for the Google Open Source Software Vulnerability Rewards Program.
Ultra-compressed communication mode. Cuts ~75% of tokens while keeping full technical accuracy by speaking like a caveman.
Comprehensive UI/UX design plugin for mobile (iOS, Android, React Native) and web applications with design systems, accessibility, and modern patterns
Creative skill for generating algorithmic and generative art. Produces visual designs using mathematical patterns, fractals, and procedural generation.
Frontend design skill for UI/UX implementation
Humanise text and remove AI writing patterns. Detects and fixes 24 AI tell-tales including inflated language, promotional tone, AI vocabulary, filler phrases, sycophantic tone, and formulaic structure.
Expert guidance for Next.js Cache Components and Partial Prerendering (PPR). Proactively activates in projects with cacheComponents: true, providing patterns for 'use cache' directive, cacheLife(), cacheTag(), cache invalidation, and parameter permutation rendering.