Loading Data

Azure Blob Storage

Store OpenSnowcat event data into Azure Blob Storage.

The easiest way to deliver high-volume, schema-validated behavioral data into Azure Blob Storage — optimized for durability, scalability, and downstream analytics.

The SnowcatCloud Loader writes enriched events directly to Azure Blob Storage in compressed TSV or flattened JSON format. Files are stored in partitioned directories, ready for use with Azure Synapse, Databricks, or custom data lake pipelines. With support for configurable flush settings and native monitoring, it's a drop-in solution for Azure-based data platforms.

Built for flexibility and performance — no infrastructure to manage, no column mapping to maintain, just clean, structured data in your Azure Blob container.

Features

  • Durable Azure-Native Storage
    Write enriched behavioral data directly to Azure Blob Storage with geo-redundant durability.
  • Partitioned Output
    Support for time-based or custom partitioning for easy integration with Synapse or Azure Data Lake processing.
  • Flexible File Formats
    Choose TSV or flattened enriched JSON, ready for querying with Azure-native tools and open-source frameworks.
  • Flush Control
    Control buffer size and flush interval to balance file frequency, latency, and cost.
  • Resilient Delivery
    Automatic retries and reliable file writing ensure no data loss during delivery.
  • Monitoring with Azure Tools
    Visibility via CloudWatch or Azure-native monitoring depending on deployment configuration.

Loader Configuration

Define how and where your enriched data is stored using a flexible configuration block:

output {
  service: "azure-blob-storage", 
  account_name: "myaccount", // xxxxx.blob.core.windows.net
  account_key: "KEYHERE",
  endpoint: "upload/prod",
  compression: "GZIP",
  format: "JSON"
},