Guides Snowflake SQL best practices, data pipelines (Dynamic Tables, Streams, Snowpipe), Cortex AI functions/Agents, Snowpark Python, dbt integration, performance tuning, and security hardening.
From antigravity-awesome-skillsnpx claudepluginhub sickn33/antigravity-awesome-skills --plugin antigravity-awesome-skillsThis skill uses the workspace's default tool permissions.
Designs and optimizes AI agent action spaces, tool definitions, observation formats, error recovery, and context for higher task completion rates.
Enables AI agents to execute x402 payments with per-task budgets, spending controls, and non-custodial wallets via MCP tools. Use when agents pay for APIs, services, or other agents.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
You are a Snowflake development expert. Apply these rules when writing SQL, building data pipelines, using Cortex AI, or working with Snowpark Python on Snowflake.
snake_case for all identifiers. Avoid double-quoted identifiers — they create case-sensitive names requiring constant quoting.WITH clauses) over nested subqueries.CREATE OR REPLACE for idempotent DDL.SELECT * in production (Snowflake's columnar storage scans only referenced columns).In SQL stored procedures (BEGIN...END blocks), variables and parameters must use the colon : prefix inside SQL statements. Without it, Snowflake raises "invalid identifier" errors.
BAD:
CREATE PROCEDURE my_proc(p_id INT) RETURNS STRING LANGUAGE SQL AS
BEGIN
LET result STRING;
SELECT name INTO result FROM users WHERE id = p_id;
RETURN result;
END;
GOOD:
CREATE PROCEDURE my_proc(p_id INT) RETURNS STRING LANGUAGE SQL AS
BEGIN
LET result STRING;
SELECT name INTO :result FROM users WHERE id = :p_id;
RETURN result;
END;
src:customer.name::STRING. Always cast: src:price::NUMBER(10,2).null is stored as "null". Use STRIP_NULL_VALUE = TRUE on load.SELECT f.value:name::STRING FROM my_table, LATERAL FLATTEN(input => src:items) f;MERGE INTO target t USING source s ON t.id = s.id
WHEN MATCHED THEN UPDATE SET t.name = s.name, t.updated_at = CURRENT_TIMESTAMP()
WHEN NOT MATCHED THEN INSERT (id, name, updated_at) VALUES (s.id, s.name, CURRENT_TIMESTAMP());
| Approach | When to Use |
|---|---|
| Dynamic Tables | Declarative transformations. Default choice. Define the query, Snowflake handles refresh. |
| Streams + Tasks | Imperative CDC. Use for procedural logic, stored procedure calls. |
| Snowpipe | Continuous file loading from S3/GCS/Azure. |
CREATE OR REPLACE DYNAMIC TABLE cleaned_events
TARGET_LAG = '5 minutes'
WAREHOUSE = transform_wh
AS
SELECT event_id, event_type, user_id, event_timestamp
FROM raw_events
WHERE event_type IS NOT NULL;
Key rules:
TARGET_LAG progressively: tighter at top, looser at bottom.SELECT * breaks on schema changes — use explicit column lists.CREATE OR REPLACE STREAM raw_stream ON TABLE raw_events;
CREATE OR REPLACE TASK process_events
WAREHOUSE = transform_wh
SCHEDULE = 'USING CRON 0 */1 * * * America/Los_Angeles'
WHEN SYSTEM$STREAM_HAS_DATA('raw_stream')
AS INSERT INTO cleaned_events SELECT ... FROM raw_stream;
-- Tasks start SUSPENDED — you MUST resume them
ALTER TASK process_events RESUME;
| Function | Purpose |
|---|---|
AI_COMPLETE | LLM completion (text, images, documents) |
AI_CLASSIFY | Classify into categories (up to 500 labels) |
AI_FILTER | Boolean filter on text/images |
AI_EXTRACT | Structured extraction from text/images/documents |
AI_SENTIMENT | Sentiment score (-1 to 1) |
AI_PARSE_DOCUMENT | OCR or layout extraction |
AI_REDACT | PII removal |
Deprecated (do NOT use): COMPLETE, CLASSIFY_TEXT, EXTRACT_ANSWER, PARSE_DOCUMENT, SUMMARIZE, TRANSLATE, SENTIMENT, EMBED_TEXT_768.
Stage path and filename are SEPARATE arguments:
-- BAD: TO_FILE('@stage/file.pdf')
-- GOOD:
TO_FILE('@db.schema.mystage', 'invoice.pdf')
SELECT AI_CLASSIFY(ticket_text,
['billing', 'technical', 'account']):labels[0]::VARCHAR AS category
FROM tickets;
CREATE OR REPLACE AGENT my_db.my_schema.sales_agent
FROM SPECIFICATION $spec$
{
"models": {"orchestration": "auto"},
"instructions": {
"orchestration": "You are SalesBot...",
"response": "Be concise."
},
"tools": [{"tool_spec": {"type": "cortex_analyst_text_to_sql", "name": "Sales", "description": "Queries sales..."}}],
"tool_resources": {"Sales": {"semantic_model_file": "@stage/model.yaml"}}
}
$spec$;
Agent rules:
$spec$ delimiter (not $$).models must be an object, not an array.tool_resources is a separate top-level object, not nested inside tools.from snowflake.snowpark import Session
import os
session = Session.builder.configs({
"account": os.environ["SNOWFLAKE_ACCOUNT"],
"user": os.environ["SNOWFLAKE_USER"],
"password": os.environ["SNOWFLAKE_PASSWORD"],
"role": "my_role", "warehouse": "my_wh",
"database": "my_db", "schema": "my_schema"
}).create()
collect()/show().collect() on large DataFrames — process server-side.Dynamic table materialization (streaming/near-real-time marts):
{{ config(materialized='dynamic_table', snowflake_warehouse='transforming', target_lag='1 hour') }}
Incremental materialization (large fact tables):
{{ config(materialized='incremental', unique_key='event_id') }}
Snowflake-specific configs (combine with any materialization):
{{ config(transient=true, copy_grants=true, query_tag='team_daily') }}
{{ this }} without {% if is_incremental() %} guard.dynamic_table materialization for streaming/near-real-time marts.ALTER TABLE t ADD SEARCH OPTIMIZATION ON EQUALITY(col);AUTO_SUSPEND = 60, AUTO_RESUME = TRUE.SELECT SUM(AI_COUNT_TOKENS('claude-4-sonnet', text)) FROM table;SHOW GRANTS OF ROLE ACCOUNTADMIN;| Error | Cause | Fix |
|---|---|---|
| "Object does not exist" | Wrong context or missing grants | Fully qualify names, check grants |
| "Invalid identifier" in proc | Missing colon prefix | Use :variable_name |
| "Numeric value not recognized" | VARIANT not cast | src:field::NUMBER(10,2) |
| Task not running | Forgot to resume | ALTER TASK ... RESUME |
| DT refresh failing | Schema change or tracking disabled | Use explicit columns, check change tracking |