Glossary/Data Storage & Compute

Compute Warehouse (e.g., Snowflake Virtual Warehouse)

Compute Warehouse is an elastic compute resource in a cloud data warehouse that allocates processing power for query execution, scaling up and down based on workload demands.

Cloud data warehouses separate storage (where data lives) from compute (processing power for queries). A compute warehouse is the processing layer: it reads data from shared storage, executes queries, aggregates results. Virtual warehouses (Snowflake's term) allow multiple compute clusters of different sizes to operate independently: small warehouse for light queries, large warehouse for heavy analytics. This separation enables cost optimization: scale storage once (data lives permanently), scale compute based on actual need (large during business hours, small at night).

Compute warehouses provide elastic scaling: add compute by scaling up (bigger virtual warehouse runs faster), add parallelism by scaling out (multiple virtual warehouses running simultaneously). Users pay for compute-seconds: a query running for 60 seconds on a 16-credit warehouse costs 16 credits; same query on an 8-credit warehouse costs 8 credits (takes longer, uses less credit). Organizations optimize by sizing warehouses appropriately: oversizing wastes credits, undersizing extends query time.

This architecture enables powerful patterns: continuous integration tests run on tiny compute warehouse (fast feedback, low cost), business dashboards run on medium warehouse (consistent performance, controlled cost), batch analytics jobs run on large warehouse when needed (fast completion, measured cost).

Key Characteristics

  • Elastic compute resources isolated from storage
  • Scales up (larger warehouse) or out (multiple warehouses)
  • Usage-based pricing: pay for compute-seconds used
  • Independent scaling from storage
  • Supports different warehouse sizes for different workloads
  • Automatic suspension to reduce idle costs

Why It Matters

  • Enables cost optimization through right-sizing for workloads
  • Provides performance isolation: one workload doesn't impact others
  • Reduces costs by not paying for idle compute
  • Enables parallel execution of independent workloads
  • Supports scaling without data re-distribution
  • Provides flexibility to scale as requirements change

Example

Snowflake analytics team: staging environment uses X-Small warehouse (1 credit/second) for development and testing, production reports use Small warehouse (2 credits/second) for consistent performance, monthly analysis jobs use Large warehouse (8 credits/second) for fast completion. Data warehouse stores terabytes of data shared across all warehouses. Different teams can operate warehouses simultaneously without interfering: marketing dashboard queries on Small warehouse don't slow finance batch jobs on Large warehouse.

Coginiti Perspective

Coginiti has deep integration with Snowflake's virtual warehouse model and similar compute-warehouse patterns on other platforms. CoginitiScript's query tags propagate metadata (department, project, priority) to Snowflake's query_tag system, enabling precise cost allocation across compute warehouses. The analytics catalog's governed, reusable logic also reduces redundant queries, helping teams right-size their compute warehouse allocation based on actual, optimized workloads rather than ad hoc query sprawl.

Related Concepts

Cloud Data WarehouseCompute ResourceElastic ScalingSeparation of Compute and StorageVirtual ClusterWorkload ManagementResource AllocationCost Optimization

See Semantic Intelligence in Action

Coginiti operationalizes business meaning across your entire data estate.