Connecting Stripe to your data warehouse or cloud storage

Stripe Data Pipeline is a no-code product that can send all your Stripe data to your data storage destination on an automated and ongoing basis.  



What data storage destinations can I connect to using Data Pipeline?

Data Pipeline supports direct integration to Snowflake and Amazon Redshift data warehouses.

It also supports connection to Google Cloud Storage, Amazon S3, and Azure Blob Storage. From there, you can send your data to various destinations, including additional data warehouses, such as Google BigQuery, Databricks, and more.

What data is available via Data Pipeline?

All Core API, Connect Core API, and Analytical datasets are available via Data Pipeline.

TheStripe Data Schema as well as our docs have a detailed view of all the data that is made available to you and how it’s organized.

How quickly is data made available in my data storage destination?

Details on data freshness are available in our docs.

How is historical data managed and loaded?

Every time fresh data is delivered to your warehouse, it will include your full dataset—from when you started on Stripe to the freshest data batch available. You are not charged for historical data.

Are there any data retention policies?

It is up to you to decide on a data retention policy and implement it. We recommend keeping data for at least seven days, but many merchants choose to only retain the last few shares to save on cloud storage costs.

Is there any sort of change data capture? How can I consume incremental data between full loads?

We do not currently support incremental data loads for our cloud storage destinations. Every table has a batch_timestamp field, which is populated with the time stamp for when data is processed through Data Pipeline.

Does Stripe perform data delivery and completeness checks?

Stripe conducts row-level checks to ensure data integrity. If any checks fail, the data won’t be delivered until complete.

Will I be notified of issues with my data delivery or planned maintenance?

When signing up for Data Pipeline, you have the option to share a contact email to receive critical updates on the status of your data processing. We suggest using a team alias so that multiple members of your team are notified about issues.

Additionally, a webhook is available to notify your system when data delivery is successful. Details on how to set up webhooks are available in our docs.

Are there any limits to the amount of data that Data Pipeline can load?

No. Data Pipeline is designed to manage large data loads. We currently accommodate some of Stripe’s largest users’ volume of 4TB of data per load. There are no rate limits.

Did this answer your question?