Skip to content

Data Pipeline: How to access

Getting access to’s Data Pipeline is easy. Using modern and secure data access techniques that are proven for scale and hosted by Amazon Web Services, you can rest assured that these access patterns are friendly to a number of open source projects and data integration tools.

Further, whether you store megabytes, gigabytes, or terabytes of data, and whether you track thousands of events per day or millions of events per day, these access patterns will work for you.

How It Works

The Data Pipeline is built upon two technologies: Amazon Kinesis for real-time data, and Amazon S3 for long-term data storage. On sign-up to the Data Pipeline service, you’ll receive access credentials to an Amazon Kinesis Stream and an Amazon S3 Bucket that we host for you.

Demoing Data Pipeline

Interested in taking a look at the data? You’re in the right place!

There are two options for you to demo the Data Pipeline Data:

  1. Download a csv file with 500 rows of sample data.
  2. See a full 30 days of Data Pipeline data by connecting to our public S3. This allows you to see exactly how the data will be deliverdd to you.

Please note that these contain sample data for the Data Pipeline product. It is intended for public use and contains no sensitive information. All data, including IPs and UUIDs, are generated artificially and not intended to be associated with any real-world person or device.

Option 1: Download a CSV

By clicking on this link, a .csv file containing 500 rows of sample data will automatically be downloaded. To understand the data in the file, please see the detailed schema page.

Option 2: Extract 30 days of data from S3

This public S3 bucket s3://parsely-dw-parse-ly-demo contains 30 days of Data Pipeline json data in zipped format (.gz). You can view the files that are available in your browser here.. To use this S3 data, please follow the steps in Getting Started to learn the different ways to connect to and extract the Data Pipeline data.

The remainder of this page is pretty technical, and goes into how, specifically, stores its raw data and provides access to that data. This page will likely only make sense to Data Scientists and programmers who have had some exposure to Amazon Web Services before. If that’s not you, but you still want to learn more, feel free to take a look at our use cases and our schema format.

Amazon Kinesis Stream

Amazon Kinesis is a hosted streaming data platform that uses to deliver our raw, enriched, unsampled streaming data to you. It is very similar to the one that uses in its own realtime data processing stream (Kafka).

The stream is similar to other “distributed message queues”, such as Kafka or Redis Pub/Sub, except you don’t have to lift a finger to set up any of the distributed infrastructure. Also, since it’s hosted by AWS, your stream is always available, replicated in three geographic regions, and “sharded” so that it can handle high data volumes without any ops work on your part.

Some technical details about Kinesis Streams set up by

  • In the us-east-1 region.
  • 1 shard per stream by default.
  • Data is retained for 24 hours.
  • Record format is plain JSON.
  • At-least-once delivery semantics.
  • Secured via an AWS Access Key ID and Secret Access Key.
  • Real-time delivery delays can be tracked at

Alternately, can deliver data to a customer-hosted Kinesis stream. This simply requires some additional setup and coordination with our support team.

Kinesis has a standard HTTP API, as well as “native” integration libraries for a number of languages. In the “code” section of this documentation, we walk through accessing a Kinesis stream using the boto3 library in Python.

Amazon S3 Bucket

Amazon S3 is a long-term, static storage engine that is globally distributed for redundancy. It is intended for more complex analysis or bulk export to different databases, like Google BigQuery or Amazon Redshift.

The S3 bucket we provide to you as part of’s Data Pipeline is similar to the bucket we use internally at for long-term storage of data, especially the backup systems behind our “Mage” time series engine. S3 allows users to pull the data and extract the information they need from it, while also giving you peace of mind that the data is safe and secure for the long-term.

Some details about your S3 Bucket with

  • In the us-east-1 region (previously called “S3 US Standard”).
  • 99.999999999% durability (“eleven 9’s”).
  • 99.99% availability (“four 9’s”).
  • Data is retained for 90 days by default, though actual retention length can vary based on contractual terms.
  • All raw events stored in the events/ prefix.
  • Data is stored as gzip-compressed plain JSON lines.
  • Typically stored in 15-minute chunks with maximum of 128 MiB of size per chunk.
  • Chunks up to 2GiB are possible for historically rebuilt/backfilled data.
  • Real-time data (from the current day) is stored, typically within the hour.
  • Secured via an AWS Access Key ID and Secret Access Key.
  • Custom one-time historical data import jobs are available upon request.

S3 also has the notion of “S3 URLs” which can be used with a variety of command-line tools.

This is our S3 bucket format:

s3://parsely-dw-mashable   # bucket name
/events            # event data prefix
/2016/06/01/12        # YYYY/MM/DD/HH
/xxxxxxxxxxxxxxx.gz      # compressed JSON file

Here is an example single S3 object:


After decompression, there would be a single JSON raw event per line, typically up 128MiB of these in a single file, which might be 10’s of thousands or 100’s of thousands of raw events.

You are not supposed to rely on the (long) .gz filename. Instead, use glob patterns or use tools that support S3 prefix or include patterns. For example:


Will match all files for the site, date 2016-06-01, and hour 12-noon UTC. During maintenance events,’s systems may populate files with a different file format in that prefix, for example all files for a single hour might be compacted into a single data.gz file in that per-hour prefix. Using a glob pattern like *.gz can guarantee you will match data even in the case of these events.

Next Steps

Further reading:

Help from our team:

  • If you are already a customer, your site is already instrumented for’s raw data pipeline. no additional integration steps are required to start leveraging your raw pageview and engaged time data. You simply get in touch with our support team or your Account Manager to add this product to your contract. From there, we will flip the switch and supply you with your secure access key ID and secret access key. From that point forward, you can use the service as you see fit, and you can also explore how to instrument custom events and custom data.
  • If you are not a customer, you simply need to go through our basic integration. Contact us for a demo and we can also walk through the integration steps necessary — which are generally pretty painless.

Due to the power and flexibility of the Data Pipeline, implementation is largely bespoke on the publisher’s end and so we cannot offer any specific implementation details here. We do, however, offer sample implementations and use cases in an open source Github repository available here, also available in Python environments via pip install parsely_raw_data.

Last updated: August 16, 2023