
Saturday, May 03, 2025
SELECT only requires read access to a customer's Snowflake account metadata database. This database only includes metadata about how the customer is using Snowflake. No actual customer data or sensitive information is stored in this database. Some examples of this information include, but are not limited to:
This data is stored in SELECT's own Snowflake account where insights can be derived and presented to our customers. We follow a principle of least privilege, and only extract the minimum subset of metadata required for SELECT's services.
We do not have read or write access to any of the customer's data that is stored in Snowflake. This access is tightly controlled during the onboarding process where customers create a new user for SELECT with an extremely limited set of permissions.
The security features discussed above can be further visualized in the diagram below outlining SELECT's secure & limited data access architecture.

See full details in the user management and RBAC documentation.
While rare, we recognize it is possible that Snowflake users within a customer's company may inadvertently include sensitive information in their query text. For example, an engineer may be debugging an issue and query all notifications sent to a particular user. They may even store some notes to themselves in the query comments:
SELECT is designed for this worst case scenario and can strip out any literal values and sensitive comments before storing the metadata in our account. This functionality can be enabled upon request. Using the same query example from above, we would only store the following query_text in our database:
Similar scrubbing can be performed across any free-form text fields ingested from the customer's Snowflake account metadata database into SELECT's database.
SELECT accesses Snowflake usage metadata to present users with insights and recommendations related to cost & performance optimization. More information on the exact views we access and their purpose is provided below.
The following views from the account_usage schema are accessed. All views contain metadata about the customer's Snowflake usage. Examples include performance statistics about historical queries run, billing amounts for different Snowflake services, and performance data for virtual warehouses. Please refer to the Snowflake documentation for each view if additional information is required. The account usage views accessed are required to present customers with comprehensive cost and performance insights.
snowflake.account_usage.access_historysnowflake.account_usage.automatic_clustering_historysnowflake.account_usage.cortex_agent_usage_historysnowflake.account_usage.cortex_aisql_usage_historysnowflake.account_usage.cortex_analyst_usage_historysnowflake.account_usage.cortex_document_processing_usage_historysnowflake.account_usage.cortex_fine_tuning_usage_historysnowflake.account_usage.cortex_functions_query_usage_historysnowflake.account_usage.cortex_provisioned_throughput_usage_historysnowflake.account_usage.cortex_rest_api_usage_historysnowflake.account_usage.cortex_search_daily_usage_historysnowflake.account_usage.cortex_search_serving_usage_historysnowflake.account_usage.database_replication_usage_historysnowflake.account_usage.database_storage_usage_historysnowflake.account_usage.document_ai_usage_historysnowflake.account_usage.materialized_view_refresh_historysnowflake.account_usage.metering_daily_historysnowflake.account_usage.metering_historysnowflake.account_usage.network_policiessnowflake.account_usage.network_rule_referencessnowflake.account_usage.network_rulessnowflake.account_usage.object_dependenciessnowflake.account_usage.pipe_usage_historysnowflake.account_usage.pipessnowflake.account_usage.proceduressnowflake.account_usage.query_historysnowflake.account_usage.search_optimization_historysnowflake.account_usage.serverless_task_historysnowflake.account_usage.sessionssnowflake.account_usage.stage_storage_usage_historysnowflake.account_usage.storage_usagesnowflake.account_usage.table_dml_historysnowflake.account_usage.table_pruning_historysnowflake.account_usage.table_storage_metricssnowflake.account_usage.tablessnowflake.account_usage.tag_referencessnowflake.account_usage.task_historysnowflake.account_usage.task_versionssnowflake.account_usage.viewssnowflake.account_usage.warehouse_events_historysnowflake.account_usage.warehouse_load_historysnowflake.account_usage.warehouse_metering_historysnowflake.account_usage.snowpark_container_services_historysnowflake.account_usage.snowflake_intelligence_usage_historyThe following views from the organization_usage schema are accessed:
snowflake.organization_usage.contract_items: Contains information about a customer's current Snowflake contract. We use this to help provide users with budgeting forecasts.snowflake.organization_usage.rate_sheet_daily: Contains information about the effective rates applied on each day. Required to calculate spend data.snowflake.organization_usage.remaining_balance_daily: Contains information about a customer's remaining contract balance. Required to determine the effective rates to apply when calculating costs and for budget forecasting.snowflake.organization_usage.usage_in_currency_daily: Contains information about how much a customer is being billed each day. Required to provide customers with Snowflake spend analytics.snowflake.organization_usage.warehouse_metering_history: Warehouse spend across the Snowflake org.SELECT accesses Databricks system tables to present users with insights and recommendations related to cost & performance optimization. Similar to Snowflake, we only access usage metadata — never actual customer data. More information on the exact system tables we access and their purpose is provided below.
The following tables from the system.billing schema are accessed to provide customers with spend analytics and budgeting forecasts:
system.billing.usage: Usage metrics by SKU, cloud, and date. Required to provide customers with Databricks spend analytics.system.billing.account_prices: Account-level pricing per SKU. Required to calculate actual spend data.system.billing.list_prices: List pricing per SKU. Used as a fallback for spend calculations.The following table from the system.query schema is accessed:
system.query.history: SQL query execution history with performance metrics such as duration, I/O, result fetch times, and caching statistics. Required to provide query performance insights.The following tables from the system.compute schema are accessed:
system.compute.warehouses: SQL warehouse configuration metadata including size, clustering, and tags.system.compute.warehouse_events: Warehouse state transitions (scaled up, scaled down, running, stopped, etc.).system.compute.clusters: All-purpose and job cluster configurations including node types, autoscaling, and runtime versions.system.compute.node_timeline: Per-node resource metrics including CPU, memory, network, and disk utilization.The following tables from the system.access schema are accessed:
system.access.audit: Audit logs used for workspace discovery and account-level metadata.system.access.table_lineage: Table-level data lineage tracking source and target tables.system.access.column_lineage: Column-level lineage tracking source and target columns.system.access.workspaces_latest: Workspace metadata including workspace name, URL, and status.The following tables from the system.lakeflow schema are accessed:
system.lakeflow.jobs: Job definitions and metadata including creator, tags, and trigger configuration.system.lakeflow.job_tasks: Task configurations within jobs including task keys and dependencies.system.lakeflow.job_run_timeline: Job execution timeline with performance metrics such as setup, queue, run, and cleanup durations.system.lakeflow.job_task_run_timeline: Task-level execution timeline within job runs.system.lakeflow.pipelines: Delta Live Tables pipeline definitions including type, settings, and configuration.The following table from the system.serving schema is accessed:
system.serving.served_entities: Served ML entity configurations including endpoint details for LLMs, external models, foundation models, and custom models.The following table from the system.storage schema is accessed:
system.storage.predictive_optimization_operations_history: Delta Lake optimization operations including compaction, vacuum, and analyze with status and metrics.