Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.streamnative.io/llms.txt

Use this file to discover all available pages before exploring further.

Preparing an external catalog is optional. StreamNative Ursa supports the following catalog modes:
  • Iceberg with the Hadoop catalog (no external catalog). Iceberg metadata and data files are written directly to the configured object storage path. No external catalog service is required, and downstream engines can read the table by pointing at the storage location.
  • Delta Lake without a catalog. Delta tables are written directly to the configured storage path; readers point at the path (s3://..., gs://..., or abfss://...) to query them.
  • External catalog. Use a managed catalog service (Databricks Unity Catalog, Snowflake Open Catalog, Snowflake Horizon Catalog, AWS S3Table, or Google BigLake) to register the tables. This is required if you want governance, discoverability, or integration with managed query engines (Databricks SQL, Snowflake, Athena, BigQuery, etc.).
If you choose to use an external catalog, this page is the index for the catalog preparation guides. Each guide focuses solely on provisioning the catalog itself and its supporting cloud resources (storage bucket, IAM role or service principal, and credentials). Once the catalog is ready, see Configure Lakehouse Catalogs for how to wire it into the compaction service.
Skip this page if you are using the Hadoop catalog (Iceberg) or no catalog (Delta). Go directly to Configure Lakehouse Catalogs, which describes the no-catalog configuration.

Catalog Support Matrix

CatalogTable FormatAWSGCPAzure
Databricks Unity Catalog (Managed Iceberg)IcebergAWSGCPAzure
Databricks Unity Catalog (Delta Lake)Delta LakeAWSGCPAzure
Snowflake Open Catalog (Polaris)IcebergAWSGCPAzure
Snowflake Horizon CatalogIcebergAWS
AWS S3TableIcebergIceberg
Google BigLakeIcebergIceberg

Choosing a Catalog

Use caseRecommended catalog
Manage Iceberg tables alongside Databricks workloadsUnity Catalog (Iceberg)
Manage Delta Lake tables alongside Databricks workloadsUnity Catalog (Delta Lake)
Cloud-portable Iceberg REST catalog from SnowflakeSnowflake Open Catalog (Polaris)
Governed Iceberg tables managed by Snowflake HorizonSnowflake Horizon Catalog
AWS-native Iceberg tables with built-in Athena/Redshift integrationAWS S3Table
GCP-native Iceberg with BigQuery/BigLake integrationGoogle BigLake

Catalog Preparation Guides

Databricks Unity Catalog (Managed Iceberg Table)

Use Unity Catalog to manage Iceberg tables governed by Databricks. The compaction service writes to Iceberg Managed Tables; downstream readers query them via the Unity Catalog Iceberg REST endpoint.

Databricks Unity Catalog (Delta Lake)

Use Unity Catalog to manage Delta Lake tables. The compaction service writes Delta Lake files governed by Unity Catalog and queryable from Databricks SQL or external Spark sessions.

Snowflake Open Catalog (Polaris)

Snowflake Open Catalog (Polaris) is a cloud-agnostic Iceberg REST catalog operated by Snowflake. It can be used as the catalog for Iceberg tables hosted on AWS, GCP, or Azure object storage.

Snowflake Horizon Catalog

Snowflake Horizon provides governed Iceberg tables with native Snowflake integration. The catalog uses a Snowflake External Volume backed by an S3 bucket and authenticates via a Programmatic Access Token (PAT).

AWS S3Table

AWS S3Table is the AWS-native Iceberg catalog with first-class integration into AWS analytics services such as Athena, Redshift, and EMR.
CloudGuide
AWSS3Table for Iceberg
Important: The Ursa cluster must run in the same region as the S3Table bucket. Cross-region access is not supported.

Google BigLake

Google BigLake provides an Iceberg REST catalog tightly integrated with BigQuery and Google Cloud Storage.
CloudGuide
GCPBigLake for Iceberg
Important: The Ursa cluster, GCS bucket, and BigLake catalog must all be in the same region. Each BigLake catalog maps to exactly one GCS bucket (1:1 mapping; sub-paths are not supported).

Next Steps

After preparing your catalog, proceed to:
  1. Dynamic Configuration Guide — Reference for all dynamic configuration keys and the cluster-name prefix requirement.
  2. Configure Lakehouse Catalogs — Connect the prepared catalog to the StreamNative Ursa compaction service.
  3. Enable Lakehouse Integration — Enable SDT (External Table) at the cluster, namespace, or topic level.