Amazon S3
Store OpenSnowcat event data into Amazon S3.
The simplest and most durable way to store high-volume, schema-validated behavioral data in Amazon S3 — ready for analytics, warehousing, or archival.
The SnowcatCloud Loader delivers enriched events directly to S3 in structured, partitioned files. Events are validated against schemas and saved in TSV or JSON formats, making them easy to query with tools like Athena, Spark, DataBricks, or Redshift Spectrum. With support for configurable buffer intervals and output formats, it's optimized for long-term storage, bulk processing, or integration into custom data workflows.
Designed for reliability and flexibility — no infrastructure to manage, no schema headaches, just clean data in your bucket.
Features
-
Durable, Cost-Efficient Storage
Store enriched behavioral data in Amazon S3 with high durability and low cost. -
Partitioned Output
Write to time-based or custom partitioned paths for efficient downstream querying and retrieval. -
Flexible File Formats
Choose between TSV and flattened enriched JSON, fully compatible with Athena and data lake tooling. -
Configurable Flush Settings
Control how often files are written based on time or size for optimal cost-performance trade-offs. -
Reliable Delivery & Retry
Built-in failure handling and retries ensure no data is lost during delivery to S3. -
Native Monitoring
Integrated with CloudWatch Metrics & Logs for visibility into loader performance and health.
Loader Configuration
Define how and where data is streamed using a simple, declarative configuration block:
output {
service: "s3",
endpoint: "s3://enriched-live-data/",
compression: "GZIP",
format: "JSON"
},