Snowflake is a cloud-native data platform that separates compute from storage and runs on AWS, Azure, and GCP. It handles SQL analytics, data engineering, machine learning, and agentic AI workloads in a single platform. As of 2026, Snowflake supports over 10,000 customers and has evolved from a cloud data warehouse into what it calls the “AI Data Cloud.”
TL;DR
- Snowflake separates compute from storage, letting teams scale each independently on AWS, Azure, or GCP.
- Current pricing starts at $2/credit (Standard), $3/credit (Enterprise), and $4/credit (Business Critical) with on-demand storage at $23/TB/month.
- Cortex AI, now generally available, brings LLM inference, agentic AI, and PII redaction directly into SQL queries.
- Full Apache Iceberg read/write support (GA October 2025) reduces vendor lock-in by letting you store data in open formats.
- Gen2 Standard Warehouses deliver approximately 2.1x faster analytics for updates, deletes, merges, and table scans.
- Snowflake Intelligence and a $200M OpenAI partnership (announced February 2026) position the platform for natural language data access.
- For teams with data in Snowflake plus MongoDB, Elasticsearch, or REST APIs, tools like Knowi can query across all sources without extraction.
Table of Contents
What is Snowflake?
Snowflake is a fully managed, cloud-native data platform built around three layers: a centralized storage layer, an independent compute layer (virtual warehouses), and a cloud services layer that handles metadata, security, and query optimization. Unlike traditional data warehouses, these layers scale independently. You can spin up multiple virtual warehouses against the same data without contention.
The platform runs natively on all three major clouds. Accounts on AWS, Azure, and GCP share the same feature set, and cross-region data sharing works without copying data. For regulated industries, Snowflake offers FedRAMP High Plus certification and Virtual Private Snowflake (VPS) for fully isolated environments.
Why has it become so successful?
Snowflake’s success comes from a combination of architectural choices, ease of use and a steady cadence of innovations.
Instant and almost Unlimited Scalability
The Snowflake architecture relies on a single elastic performance engine that offers high speed and scalability. Snowflake can also support a high number of concurrent users and workloads, both interactive and batch. This can be attributed to its multi-cluster resource isolation feature. Snowflake is robust and high-performing, giving organizations the confidence they need to handle all types of data workloads.
In June 2025 Snowflake introduced the Standard Warehouse – Generation 2 (Gen2), which delivers roughly 2.1× faster price/performance for analytics compared with the earlier generation. A new Adaptive Compute service (private preview) selects cluster sizes, number of clusters and auto‑suspend/resume durations automatically, further optimising cost and performance.
Unified Data Processing
Snowflake users can query structured, semi‑structured and (with recent releases) unstructured data using standard SQL. Features like Document AI, powered by Snowflake’s Arctic‑TILT large language model, convert PDFs, invoices and other unstructured documents into structured tables. Document AI became generally available in late 2024 across most AWS and Azure regions.
A Single Data Copy, Shared Securely
Snowflake’s cross‑region and cross‑cloud sharing means teams work from one consistent copy of data. The Snowflake Horizon Catalog federates across Apache Iceberg tables through catalog‑linked databases, giving users unified governance and discovery across Snowflake and external sources. Horizon Catalog now includes a copilot that uses natural language to assist with governance and metadata discovery tasks.
FinOps and cost visibility
Snowflake joined the FinOps Foundation in 2025 and rolled out features to help customers monitor and control spending. These include Cost‑Based Anomaly Detection (public preview) for spotting abnormal cost spikes, Tag‑Based Budgets for allocating spend to teams, and tools such as Performance Explorer, Query Insights and Cost Insights for Large Query Gaps.
Security and governance
Recent releases added leaked‑password protection and Bad IP Protection to disable compromised credentials. Snowflake expanded Trust Center detectors and introduced new MFA methods, including passkeys, Windows Hello and security‑key support. It also added personal access tokens and workload identity federation to replace long‑lived passwords, plus sensitive‑data monitoring, synthetic data generation and enhanced Private Link connections.
Business continuity
Snowflake’s snapshots (public preview) create point‑in‑time, immutable backups that cannot be modified or deleted, even by administrators, helping organisations meet regulatory and ransomware‑protection requirements. Improved replication interfaces and schema‑level replication make disaster recovery more intuitive and cost‑effective.
Extensibility and ecosystem
Snowpark Container Services (SPCS) lets customers run containerised workloads (e.g., AI model training, custom applications) directly inside Snowflake, with general availability on GCP expected soon. In development is Snowflake Postgres, a fully managed PostgreSQL service within Snowflake. Unistore, Snowflake’s transactional engine, expanded to Azure with Hybrid Tables and now offers Tri‑Secret Secure (TSS) encryption.
Dynamic tables enhancements
Snowflake’s Dynamic Tables simplify data pipelines by refreshing downstream tables automatically.
- Filtering by current date/time, allowing SQL filters to minimise processing costs.
- Immutability, letting users mark regions of a dynamic table as read‑only to preserve historical data and avoid costly recomputation.
- Insert‑only inputs, which process only new inserts for faster performance.
- Backfill, enabling dynamic tables to be seeded from existing data using zero‑copy cloning.
What Snowflake Handles in 2026
Snowflake started as a cloud data warehouse. It now covers six workload categories:
- Data warehousing: SQL analytics on structured data, which remains the core use case.
- Data engineering: ETL/ELT pipelines with Dynamic Tables, Snowpipe Streaming, and the new Openflow ingestion service (built on Apache NiFi).
- Data science and ML: Snowpark for Python/Scala/Java, ML Jobs for distributed training, and Snowflake Notebooks on a Jupyter kernel.
- Data applications: Build and distribute apps via Native Apps Framework, Streamlit, and Snowpark Container Services.
- Data sharing: Zero-copy sharing across accounts, Snowflake Marketplace, and Data Clean Rooms for privacy-preserving collaboration.
- AI and agentic workloads: Cortex AI functions in SQL, Cortex Agents for autonomous task execution, and Snowflake Intelligence for natural language querying.
Key Snowflake Features (2025-2026 Updates)
Cortex AI: SQL-Native Intelligence
Cortex AI is Snowflake’s AI layer, letting you run LLM inference, sentiment analysis, transcription, and PII redaction directly in SQL queries. No external API calls, no data leaving Snowflake. Key Cortex features and their current status:
- AI_COMPLETE: GA (November 2025). Run LLM inference with structured JSON outputs.
- Cortex Agents: GA (November 2025). Build autonomous AI applications that retrieve data from structured and unstructured sources.
- Cortex Code: GA (February 2026). AI coding agent trained on Snowflake’s own engineering practices. Automates pipeline building, governance policies, and access controls.
- Cortex Fine-tuning: GA (February 2025). Train custom models on your data within Snowflake.
- AI_REDACT: GA (December 2025). Automated PII detection and redaction.
- AI_TRANSCRIBE: GA (November 2025). Audio-to-text transcription.
- AISQL: Preview (June 2025). Multimodal AI processing, including images and documents, within SQL.
Snowflake Intelligence and the OpenAI Partnership
Snowflake Intelligence (preview, August 2025) lets users query enterprise data using natural language. It combines Cortex Agents with LLMs from Anthropic, OpenAI, Meta, and Mistral to answer questions across both structured tables and unstructured documents.
At Build 2026, Snowflake announced a $200 million partnership with OpenAI, bringing frontier models directly into the Intelligence interface. Semantic View Autopilot, also announced at Build, uses query history to auto-generate semantic views, reducing manual data modeling.
Apache Iceberg and Open Formats
Snowflake’s biggest move against vendor lock-in. As of October 2025, Iceberg table support is fully GA for reads, writes, partitioned writes, and row-level deletes. You can now store data in open Apache Iceberg format while still using Snowflake’s compute and governance.
- Catalog-linked databases: GA. Connect Snowflake to external Iceberg catalogs.
- Iceberg replication: GA. Failover and failback support for Iceberg tables.
- pg_lake: GA. Open source PostgreSQL extension that manages an Iceberg catalog.
- Cross-format sharing: Share Iceberg and Delta Lake tables via zero-ETL sharing.
Data Engineering Updates
- Gen2 Standard Warehouses: GA (May 2025). Approximately 2.1x faster for updates, deletes, merges, and table scans.
- Dynamic Tables with timestamp filtering: GA (May 2025). Time-based incremental refresh.
- Snowflake Openflow: GA (June 2025). Managed ingestion service built on Apache NiFi with hundreds of connectors for ETL and CDC.
- dbt Projects on Snowflake: GA (November 2025). Native dbt development and execution with Git integration in Snowsight.
- Snowpipe Streaming (high-performance): GA (September 2025). Enhanced real-time data ingestion.
Developer Tools and Applications
- Snowpark Container Services: GA on AWS and GCP. Run custom Docker containers, ML models, and full-stack apps.
- Snowflake Notebooks: GA. Built on a Jupyter kernel for end-to-end data science workflows.
- Shared Workspaces: GA (September 2025). Collaborative development environments with role-based security.
- Snowflake Postgres: Preview (December 2025). Fully managed PostgreSQL compatibility layer within Snowflake.
- Native Apps Framework: GA. Build, distribute, and monetize apps through Snowflake Marketplace.
Governance and Security
- Horizon Catalog: GA. Unified governance with cross-catalog discovery across Iceberg and relational tables.
- Data Quality Expectations: GA (August 2025). Automated monitoring framework.
- WORM Backups: GA (December 2025). Immutable backups that cannot be modified or deleted, even by administrators.
- Cost Anomalies: GA (December 2025). Automated spending deviation detection.
- Process Lineage: GA (October 2025). End-to-end tracking across tasks and stored procedures.
Snowflake Pricing in 2026
Snowflake uses a consumption-based model. You pay separately for compute (credits per second of warehouse uptime) and storage (compressed TB per month). Current pricing for AWS US East (Northern Virginia), sourced from Snowflake’s pricing page:
| Edition | Price per Credit | Key Features | Best For |
|---|---|---|---|
| Standard | $2.00 | Core platform, encryption, Snowpark, data sharing, Time Travel | Small teams getting started, development workloads |
| Enterprise (Most Popular) | $3.00 | + Multi-cluster compute, granular governance, extended Time Travel | Growing teams needing concurrency and governance controls |
| Business Critical | $4.00 | + Tri-Secret Secure, private connectivity, failover/failback | Regulated industries (healthcare, financial services) |
| Virtual Private Snowflake | Contact sales | Completely isolated Snowflake environment | Government, defense, highest security requirements |
Storage: $23.00 per TB/month (on-demand, compressed). Pre-purchased capacity discounts available.
How costs scale: Warehouse sizes range from X-Small to 6X-Large. Larger warehouses consume more credits per second but process queries faster. Serverless features like Cortex AI, Document AI, and Snowpipe Streaming have their own credit consumption rates documented in Snowflake’s Consumption Table.
Two purchase options: On-demand (pay full price, no commitment) or Capacity (pre-purchase credits at a discount, best for predictable workloads). Snowflake offers a 30-day free trial with no credit card required.
Compute (Virtual Warehouses & Credits)
Credits are the billing unit for compute. Every running warehouse consumes credits by the second (with a minimum of 1 minute), even when idle. The number of credits per second depends on the warehouse size (X‑Small through 6X‑Large).
Example warehouse sizes:
| Size | Credits/Hour | Credits/Second |
|---|---|---|
| X-Small | 1 | ~0.0003 |
| Medium | 4 | ~0.0011 |
| Large | 8 | ~0.0022 |
| … | … | … |
- Credit costs vary by edition and region: Standard is roughly $2-3/credit, Enterprise $3-4.65, Business Critical $4-6.20, and VPS $6-9.30+ in U.S. regions .
- Cloud Services: This covers metadata, optimization, and coordination workloads. Snowflake offsets this by billing a maximum of 10% of your compute credits per day, whichever is lower .
Storage
- Billed as a flat rate per TB per month, based on the compressed data size stored—including Time Travel history and staged data. Snowflake optimizes compression automatically.
- The exact cost depends on region and on whether you’re using On-Demand or Capacity (pre-purchased) plans.
- Storage fees are generally a smaller portion of total cost, often under 10%, since compute typically dominates 80–90% of your bill.
Data Transfer
- Inbound data (uploads): Free, no charges.
- Outbound data (egress): Charged based on region and cloud provider. Transfer within the same region/cloud is free; cross-region or cross-cloud transfers are billed per byte.
Purchasing Options
- On-Demand: No upfront commitment, pay full price per credit—best for flexible or variable usage.
- Capacity (Pre-Purchased): Commitment-based credits can significantly reduce compute and storage costs—ideal for steady, predictable workloads.
Summary Table
| Cost Component | Billing Unit & Behavior | Influenced By |
|---|---|---|
| Compute | Credits per second (warehouse size × running time) | Edition, region, warehouse size, usage patterns |
| Cloud Services | Daily adjustment up to 10% of compute credits | Overall compute usage |
| Storage | Flat rate per compressed TB/month | On-Demand vs Capacity, region |
| Data Transfer | Per-byte egress charges | Cloud provider, regions involved |
| Capacity vs On-Demand | Commitment-based discounts vs flexibility | Forecasted usage stability |
Snowflake vs Competitors (2026 Comparison)
The cloud data platform market has four major players in 2026. All now support SQL, Python, and AI workloads. The decision comes down to your cloud strategy, team skills, and AI roadmap.
| Dimension | Snowflake | Databricks | BigQuery | Microsoft Fabric |
|---|---|---|---|---|
| Core Strength | SQL analytics, data sharing, ease of use | Data engineering, ML/AI, open source ecosystem | Serverless analytics, zero infrastructure management | Unified Microsoft stack with native Power BI |
| Architecture | Multi-cloud SaaS, separate compute/storage | Open lakehouse (Delta Lake, Apache Spark) | Fully serverless, auto-scaling | Azure-native, OneLake unified data lake |
| Multi-Cloud | AWS, Azure, GCP (full parity) | AWS, Azure, GCP | GCP-native (Omni for multi-cloud reads) | Azure only |
| AI/ML Approach | Cortex AI: SQL-native LLM functions, agents | Mosaic ML, MLflow, Unity Catalog, open models | Gemini integration, BigQuery ML | Azure AI, Copilot, Azure ML |
| Pricing Model | Credits per second of compute | Databricks Units (DBUs) | Per TB scanned ($6.25 on-demand) or slots | Capacity Units (CU) |
| Open Format Support | Apache Iceberg (GA read/write), pg_lake | Delta Lake native, Iceberg interop | BigLake for open formats | OneLake with Delta/Parquet |
| Best For | Multi-cloud enterprises, data sharing, SQL-first teams | Data engineers, ML teams, open-source preference | Google Cloud shops, variable ad-hoc workloads | Microsoft-first organizations, Power BI users |
Snowflake vs Databricks
Databricks leads for heavy ML workloads and data engineering teams that prefer open source tools. Snowflake leads for SQL-first analytics, data sharing across organizations, and teams that want a managed experience without infrastructure management. Snowflake’s Iceberg support narrows the lock-in gap, but Databricks’ Delta Lake remains more mature for lakehouse architectures.
Snowflake vs BigQuery
BigQuery’s fully serverless model means near-zero cold start times and no warehouse management. Snowflake offers more control over compute resources and better multi-cloud support. BigQuery’s per-query pricing works well for variable workloads but can get expensive for heavy, consistent querying where Snowflake’s per-second billing is more predictable.
Snowflake vs Microsoft Fabric
Fabric’s strength is its integration with the Microsoft ecosystem. Direct Lake mode lets Power BI read from OneLake without imports, and 70% of the Fortune 500 already use Fabric through their Microsoft contracts. Snowflake wins on multi-cloud portability and cross-organization data sharing. If your stack is Microsoft-first, Fabric reduces friction. If you need multi-cloud or complex data sharing, Snowflake is the safer choice.
Snowflake Limitations to Consider
Snowflake is powerful, but it is not the right fit for every workload. Five things to evaluate before committing:
- Cost escalation: Credit consumption grows quickly with larger warehouses, Cortex AI features, and serverless compute. Active cost management is essential. The Cost Anomalies feature (GA December 2025) helps, but does not replace governance.
- Vendor lock-in: Historically high with Snowflake’s proprietary format. Iceberg support (GA) and pg_lake mitigate this, but most existing customers still use Snowflake-native tables.
- Streaming workloads: Snowpipe Streaming and Dynamic Tables have improved real-time capabilities, but Kafka and Flink remain better for sub-second streaming requirements.
- AI maturity: Cortex AI features reached GA in late 2025 and early 2026. For teams doing heavy custom ML training, Databricks’ ecosystem (MLflow, Mosaic ML) is more established.
- Optimization complexity: Warehouse sizing, clustering keys, and materialized views require expertise to get right. Gen2 warehouses and Adaptive Compute (preview) help, but Snowflake is not fully self-tuning yet.
Analyzing Snowflake Data with Knowi
Snowflake handles storage and compute well, but many teams also have data in MongoDB, Elasticsearch, REST APIs, or other sources alongside Snowflake. Building dashboards that combine these sources typically requires extracting everything into one warehouse first.
Knowi connects directly to Snowflake and queries data in place, with no extraction or staging layer. It also connects natively to MongoDB, Elasticsearch, Cassandra, DynamoDB, and REST APIs, allowing cross-source joins in a single query without moving data.
- Direct Snowflake connection: Query Snowflake data as-is. No BI Connector overhead, no pre-modeling.
- Cross-source joins: Combine Snowflake tables with MongoDB collections or REST API responses in one query.
- Embedded analytics: White-label Snowflake dashboards into your application with multi-tenant row-level security.
- Natural language queries: Non-technical users can ask questions in plain English across Snowflake data without learning SQL.
For a step-by-step walkthrough, see the Snowflake analytics tutorial.
Frequently Asked Question
What is Snowflake and how does it differ from a traditional data warehouse?
Snowflake is a cloud-native data platform that separates compute from storage, runs on AWS, Azure, and GCP, and scales each layer independently. Traditional on-premises warehouses couple compute and storage, requiring capacity planning and hardware management. Snowflake eliminates this by letting you spin up virtual warehouses on demand and pay per second of usage.
How much does Snowflake cost in 2026?
Snowflake pricing starts at $2.00 per credit for Standard edition, $3.00 for Enterprise, and $4.00 for Business Critical (AWS US East). Storage costs $23.00 per compressed TB per month on-demand. Actual costs depend on warehouse size, query volume, and usage of serverless features like Cortex AI. Pre-purchasing capacity credits lowers the effective per-credit price.
What is Cortex AI in Snowflake?
Cortex AI is Snowflake’s built-in AI layer that lets you run LLM inference, sentiment analysis, text summarization, PII redaction, and audio transcription directly in SQL queries. It reached general availability in November 2025 and supports models from Anthropic, OpenAI, Meta, and Mistral. Cortex Agents (also GA) enable building autonomous AI applications within Snowflake’s governance framework.
Does Snowflake support Apache Iceberg?
Yes. As of October 2025, Snowflake supports full read and write operations on Apache Iceberg tables, including partitioned writes, row-level deletes, and cross-account replication. Catalog-linked databases connect Snowflake to external Iceberg catalogs. This lets you store data in open formats while using Snowflake’s compute and governance, reducing vendor lock-in.
How does Snowflake compare to Databricks in 2026?
Snowflake is strongest for SQL analytics, data sharing, and managed simplicity. Databricks leads for data engineering, ML model training, and teams that prefer open source tools like Delta Lake and MLflow. Both platforms now support SQL, Python, and AI workloads. The choice depends on whether your team is SQL-first (Snowflake) or code-first (Databricks).
Can you run AI models inside Snowflake?
Yes. Cortex AI functions run LLM inference directly in SQL without external API calls. Snowpark Container Services lets you deploy custom Docker containers with your own models. ML Jobs (GA August 2025) support distributed model training. For teams that want AI without data leaving Snowflake, Cortex provides a governed, SQL-native approach.
Can you query Snowflake data alongside MongoDB or REST APIs without ETL?
Not natively within Snowflake. Snowflake connects to its own storage layer. However, platforms like Knowi connect to Snowflake, MongoDB, Elasticsearch, and REST APIs simultaneously, enabling cross-source joins and dashboards without extracting data into a single warehouse.