We are happy to announce the ability to push data to AWS S3 using our data push feature. This adds another option when building your own data warehouse/data lake.
Part of our mission is to offer simple ways to take your social media data out of quintly to avoid data silos and enable analysis across several sources. Our data push feature handles scenarios where you want to build your data warehouse/data lake. Our system proactively pushes structured data to you on a defined schedule, eliminating extra work on your end.
Our API provides a generic interface to take data out of quintly. The data push feature is an extra layer on top of our API. It takes away all of the development work required to connect to and maintain quintly's API. Once activated, it will proactively push the data you need to your data warehouse/data lake on a defined schedule. We make sure nothing goes wrong and provide transparency into each of the data push tasks performed.
If you want to analyze data on the spot and are not interested in building your own data warehouse/data lake, recently we have expanded our integrations library by adding a native Google Data Studio connector.
Both approaches have pros and cons, but in the true nature of flexibility, we strive to cover each use case.
With this update, we can push data to an AWS S3 bucket and a Google BigQuery database, covering two popular services offered by AWS and Google Cloud. We look forward to add more destinations very soon.
If you are interested in pushing data to your data warehouse/data lake, please contact our customer service at support@quintly.com or via chat in-app.