quintly Product changes logo

Product changes

Back to Homepage Subscribe to Updates

Labels

  • All Posts
  • New feature
  • Improvement
  • Data
  • Fix

Jump to Month

  • February 2023
  • December 2022
  • October 2022
  • September 2022
  • August 2022
  • July 2022
  • June 2022
  • May 2022
  • April 2022
  • March 2022
  • February 2022
  • January 2022
  • December 2021
  • November 2021
  • September 2021
  • August 2021
  • July 2021
  • June 2021
  • May 2021
  • April 2021
  • March 2021
  • January 2021
  • December 2020
  • November 2020
  • October 2020
  • September 2020
  • August 2020
  • July 2020
  • June 2020
  • April 2020
  • March 2020
  • February 2020
  • January 2020
  • December 2019
  • November 2019
  • October 2019
  • September 2019
  • August 2019
  • July 2019
  • June 2019
  • May 2019
  • April 2019
  • March 2019
  • February 2019
  • January 2019
  • December 2018
  • November 2018
  • October 2018
  • September 2018
  • August 2018
  • July 2018
  • June 2018
Improvement
3 months ago

External profile addition and authentication - the power to end user 💪

We are thrilled to deliver a new way of handling profile addition - external flow involving the end account owner. It is a welcomed addition for agency-type setups where quintly's account needs to analyze multiple externally managed profiles. Security and proactivity in the center, let's dive in.

 

 

Why is it important?

Profile addition was always a complicated use case when handling externally owned accounts, and it required loads of communication, reminders, and… in the worst-case scenario, password sharing. One of the most prominent client types for quintly is Agencies. Due to the nature of their business, analytics of externally owned accounts is a must. A few years ago, we implemented an authentication process to handle that, where the user can share a link with the external user, user logs in, and authenticate selected profiles, once they are added. We have decided to expand on this functionality and create a flow for the end client to add and authenticate their profiles. It's essential for networks where profile search, due to the technical API setup, is impossible ( TikTok and Snapchat Shows). 

How does this flow work?

To request the end user add and authenticate accounts, go to the Accounts > Profiles section and click “New profile”. Underneath the "Continue with…" Social network buttons, you will find a completely new button, "Copy authentication link". Once clicked, you can share copy, saved in your local clipboard, with the account owners. Upon opening the link, they will gain access to the personalized landing page. From here, they can use the native application flows to add and authenticate any accounts they manage. 

Once finished, you will receive a notification via email and our in-app notification center that the profiles are ready to use. You can find them in the profile addition modal under the "Profiles connected with authentication link" section. You can now follow the standard profile addition. Bonus point - you already have access to private data! 

Give control, and gain data.

What has started as a necessity (network limitation), in a true quintly's fashion, became a full-fledged feature. It fits in perfectly with the agency's use case, and we cannot wait for you to take advantage of this addition fully. Less communication, more data, saved time - we support that!

ImprovementData
5 months ago

Data Push Tasks - a true NO CODE solution for your social media data

We are thrilled to bring forth one of our most powerful features - Data Push Tasks. For the first time, you will have complete control over what data you would like to push to your designated data warehousing solution. The best part? No development effort on your side! Let’s dive in.

  

What are Data Push Tasks

Here at quintly, flexibility is one of our core beliefs, and it also translates to extending quintly into the whole MarTech ecosystem. That is why we have introduced three ways to help users get our quality-driven data into their IT infrastructure. Depending on the tech stack, our customers can choose the following options: get data via API, native integrations (google data studio and Tableau) or Push Tasks. 

In a nutshell, Push tasks are scheduled events that would send all of your previously selected social media data to the data warehousing solution of your choice. The data would be quality-checked before transfer. Once your data is available, you can add it to your marketing stack.

How to get started

Please visit our integrations / Data Push Tasks page to initialize a push task. In this section, you can add new tasks, view your already scheduled ones or consult the limits. 

When adding a new task, you must define which social network you want to use, which data source, and which columns are relevant. Choose your data provider (we support Google Big Query and Amazon S3), authenticate, and finish the setup with advanced (CRON-based) scheduling, and you are ready to go. 

Save money on development

The most significant advantage of the Data Push Tasks is the “No Code” approach to the topic. All you have to do is to set up the task, and we will handle the rest for you, delivering the best quality data into your warehousing solution. We wanted to ensure that even if you do not have an extensive development budget, you can connect quintly’s data to your data insights tech stack. We offer this service at a fixed cost of €300 per social network, not depending on the data size. 

Is this product for you

Whether you are thinking about bringing social media data to your warehousing solution or you have access to limited development resources, the answer is - absolutely. It’s a great extension built on a solid foundation of quinlty’s data. If you are interested to find out more, contact our support team.

Improvement
7 months ago

Improved custom notes on metrics

Recently we have updated a metric component, custom notes, that allows users to leave metric-specific messages visible on the dashboard level. Thanks to this improvement, we have made the storytelling aspect of a single metric more accessible. Let’s dive in.

Describe what you see

To add a note to the single metric, click on the “+Notes” button in the footer of the metric. An additional popup with a simple WYSIWYG editor will show up. Once you have created your metric, click “Save”, and the newly created custom note will appear in the footer. You can always modify or delete them if you wish

Publish and deliver your personalized findings

The most beneficial aspect of improved custom notes is in metric and dashboard sharing. Custom metric messages are widely supported by all our sharing (interactive reports) and export options (image formats, pdf, excel, CVS, and Powerpoint). Thanks to this inclusion, data analysts can attach their insights or descriptions on the metric level to anyone who is the end user. I hope that you will find this update helpful.

ImprovementData
8 months ago

Interactive reports now available via API

A sought-after update to our API will allow you to create share links via new endpoints. A very useful feature for our clients/partners integrating with quintly via API


Shared interactive report links

One of our core shared features allows users to share dashboards in an interactive format. We always believed that when it comes to data, it is crucial to give the end viewer ability to zoom in on the information they are looking at. App users were able to take advantage of managing the links for quite a while, so we are happy that our API users can do it programmatically now too.


New API endpoints

To manage the dashboard share links, we made four new endpoints available: listing all shared links, creating, updating, and deleting share links. Creating and updating come with parameters, including timeframes, password protection, and expiration options. You can find the whole list of features in our API documentation. We have more exciting news coming, so stay tuned 🙇 .

Improvement
9 months ago

New inline grouping and tagging

We are happy to bring you a new powerful UI component - Inline grouping and tagging. This universal UI module will allow you to tag and group profiles on the fly, massively increasing your productivity within the tool.

  

Get organized

The inline component completely replaces the old mechanism of adding, removing, and modifying groups and tags. To access this component, click on the pencil icon next to profiles or posts. This will open a dialog window where you can quickly search for existing groups or tags and modify or remove them. We have unified the UI of this component for a more streamlined experience. You can still access and manage tags and groups in their respective sections within our platform.

Components

We have spent quite a bit of time reviewing the most used components and challenging how they are used. The goal of this exercise was always an improvement in how you will use our platform. Stay tuned for more updates.

Improvement
10 months ago

Q2 smaller updates

📅  Dynamic date ranges.

We bring you a date selector on the dashboard level. We have made it available to choose from two types of date settings - fixed and rolling date. For a set timeframe - fixed is perfect. A rolling date range would be better if you want a more dynamic date range. Once set, the dashboard will remember this setting. Dynamic date ranges are available for interactive and scheduled reports.

😎  Improved onboarding flow.

We have simplified and improved our onboarding flow, allowing you to add profiles and authenticate them in quick steps. You are also able to invite new users directly from this flow. More streamlined, more self-serve, and will get you up and running in minutes :)

😀  Linkedin new reaction.

As you probably know, Linkedin has rolled out the “Funny” reaction. We have it implemented immediately, so this new reaction type is available in LinkedIn's Posts Table default metric.

📌  Data source section.

Minor update to our data source section: The column types (INTEGER, DATETIME, etc.) now link to a knowledge base article describing the data types in more detail.

💲 Ad break metrics for facebook.

We have added new ad break metrics for individual Facebook posts and videos from Insights. Ad break metrics cover impressions, earnings (in USD), and CPM (earnings per 1000 impressions) for advertisement in a Page's videos. These are the metrics.

For posts (Definition):

  • post_video_ad_break_earnings
  • post_video_ad_break_ad_impressions
  • post_video_ad_break_ad_cpm

For videos (Definition at the bottom of the page):

  • total_video_ad_break_earnings
  • total_video_ad_break_ad_impressions
  • Total_video_ad_break_ad_cpm

Those metrics have been added to our facebookInsightsOwnPosts and facebookInsightsOwnVideos, respectively. They are ready to use in custom metric builder and via API.

📥  Data Push tasks improvements.

We have made improvements to our data push engine. You can now access your Data Push tasks configuration using our API with a new endpoint. 

🔩  CSV importer for Migration assistant

We have finished an improved migration assistant script. It allows us to migrate public fan, follower, and subscriber counts from an old tool into quintly in various CSV formats.

ImprovementData
10 months ago

Data Quality and Performance Improvements

Over the last couple of months, our team has been working on a massive data structure and process overhaul. The main task what to replace the last remaining real-time data sources. Why is it a big deal? Let's dive in.

Real-time VS persisted data sources.

Based on some data retrieved from social networks, a real-time API call was needed. A solution that worked quite well, but unfortunately, due to the volume of data, it was slow, we couldn't pass the data through our Data Layer in real-time, and building custom metrics was quite complicated due to the response format (JSON). It was especially problematic if there was an issue with the information delivery. Due to the real-time nature, it was tough to debug the origin of the case. We have decided to move the remaining real-time data sources within our Data Layer infrastructure, with the help of our import workers, effectively changing them to persisted data. Persisted what?

To keep it short, this type of information has been imported via our workers, treated by our Data Layer engine, and stored in our database, becoming easily accessible and queryable.

Data Layer

One of the pillars of quintly, Data Layer, is more than just a database storing information. It's a sophisticated and constantly evolving data curation engine to ensure that the information served to our clients is the most accurate and qualitative. Thanks to moving all of the remaining data sources within our Data Layer, we have achieved:

  • Robustness: thanks to structuring the information the same way, you are free to combine the data sources the way you want with ease 
  • Performance improvement, no more ad hoc parsers, we query already sanitized data from one source.
  • Transparency: since we control the data on our end, we can track the whole journey and report on it better
  • Data quality and consistency: as mentioned above, each data source will undergo treatment from Data Layer, performing various data quality checks and improvements

This iteration is also a huge deal for us internally. It will significantly reduce the time needed for the maintenance, bug tracking, and, most importantly, implementation of the new data sources.

Writing about the data quality and maintenance work involved is very hard. Still, this improvement is significant as it helps us continue providing the most sanitized data for all your analytical needs. Isn't that what we are all here for anyway?

ImprovementDataFix
a year ago

Q1 2022 smaller updates ☃


The first quarter of 2022 brought a lot of bigger product updates mentioned here before. We have also worked on the more minor updates related to either iteration of feature release, customer feedback, or our discovery. We have put all of them below. 


Updates to our Data Push and API

👉 We have made it possible to fetch all of the metadata of all your data push tasks via /list-data-push-tasks API endpoint. It's the same information as the one from within our app in the Integrations > Data Push section. You can find more information in our API documentation.

👉 We also support pushing JSON format to AWS S3, supplementing the already existing CSV format. You can find information about the data format on the data push task detail page.

👉 One more minor update to those using our API and data push tasks. We will now return the primary key per data source. This is especially useful for data warehouses, where you would like to understand what set of data source columns identifies a row within.

Data sources and general improvements

👉 Small but powerful update to the profiles data source. It now includes a column containing the complete set of groups to which the profile belongs. It enables metrics that want to aggregate results per group and thus allows groups to be used as a kind of "label" for profiles.

👉 The expiry date is no longer mandatory when generating a new interactive link. You can share and access the interactive links indefinitely. 

👉 Finally, if you create a link for someone to authenticate the profile, you will be notified via email once the task has been completed.

Q2 and beyond

We hope that you liked this format and we will continue bringing more of those smaller updates to light in a quarterly fashion. We have quite a few big updates coming up too so watch this space.

ImprovementData
a year ago

Filter support for time-series data 💪

We are happy to bring a significant data overhaul that we have done to time-series data sources. We give you fully accessible, per post, and filterable time-series data. Why is this huge? Let's find out.

  

Humble beginnings

The sheer amount of data queried in our platform at each given moment creates challenges when it comes to the loading times and performance. The time-series data was most affected, so we have decided to propose a solution - aggregation tables. In a nutshell, instead of getting information on the posts, we stored the critical metrics with timestamps in those aggregation tables. This was a perfect solution for many years, but when there is progress, there are welcomed changes and challenges - we have introduced powerful post tags and filters.   

The aggregation tables were a great solution at the time. Still, with the introduction of tags and filters, it was impossible to add this functionality to those tables as the aggregation concept works on the number of posts and key metrics and not on the individual posts themselves. That meant that metrics based on aggregation tables could not take advantage of the advanced post filtering when needed. 

Over the last few months, we have improved our data infrastructure, which allowed us to revisit the topic of how to support filters for time-series data. Today we can happily say that we finalized it.  

What does it mean to you?

Simply put, any metric based on post data can now be further filtered down by our advanced filters in real-time. Would you like to see how tagged posts, hashtags or mentions did over time? You have a social media campaign running, and you want to do competitive benchmarking on top-level KPIs? Now you can. In terms of functionality, nothing changes. You can still use and build the same metrics, with the difference that popular metrics like "Key Metrics Table" and "Number of Posts", if there are filters applied, the total aggregated results will show only the filtered posts where requested. This change will allow you to build even more robust dashboards that can start with the most generic KPIs, drilling down to post level to tell the full zoomed-in story. It is worth mentioning that this also applies to our auto-tagging filter.

There are a few small prints that we need to list below.

  • This change only applies to post related metrics as it operates on the post titles and descriptions, so followers and user metrics won't be affected.
  • Filtering by tag won't be available for User Generated Content (members posts on your timeline), but the post text filter will. Our system will recognize this for you and present it accordingly.
  • Due to the sheer amount of data that has to be parsed ad hoc, in a very rare scenario (we are talking hundreds of profiles with thousands of posts) that the generation of the result might take more than 20 seconds

What the future holds

This update was an important missing piece to provide you with genuinely consistent data. It also standardizes all the processes within quintly and removes unnecessary maintenance. Our advanced filters will only get more powerful over time (we have a few things planned), so this alignment is really welcomed. Most importantly, this improvement allows you, our users, to zoom in and filter out the noise where it matters the most. We cannot wait to see what you come up with and how you will use it. That's our biggest win.

ImprovementData
a year ago

Twitter polls now available ✅

We are happy to announce that you are now able to analyze Twitter polls. This is a welcome addition to the already extensive Twitter data sources. Let’s dive in.

Opinions as data metric

In Twitter’s own words: “Twitter Polls allow you to weigh in on questions posed by other people on Twitter. You can also easily create your own poll and see the results instantly”. Polls have been heavily used by brands and media companies alike and are the best source of instant question validation. We have decided to include polls as a part of our private-level twitterAnalyticsOwnTweets data source. We have added a few columns, including:

Column nameTypeDescription
isPollBOOLEANWhether this Tweet has a poll attached or not.
pollIdSTRING>The ID to the attached poll.
pollDurationMinutesINTEGERNumber of minutes the poll attached to this Tweet lasted.
pollEndTimeDATETIMEThe time the poll attached to this Tweet ended.
isPollClosedBOOLEANWhether the poll attached to this Tweet is still active and can receive votes, or if the voting is closed.
pollVotesINTEGERNumber of votes the poll attached to this Tweet got across all choices.
pollOptionsOBJECTThe choices of the poll attached to this Tweet, including the vote count on each choice. Provided as a list of objects, e.g. [{"pos": 1, "label": "Choice A", "votes": 50}, ...].

In a true quintly fashion, we have accompanied this new addition with new default metrics (chart and table) to get you started.

What’s next for Twitter?

We will continue to work on adding those extra data points to our data sources. Our private-level data sources are already running on Twitter V2 API, but we are working on moving all public data requests to the new version as well. Not only is it the newest, most supported version, but also comes with brand new metadata. This change will enable new metrics currently not available, exciting 😎 .