Users can now choose to expand the results of path analysis by a particular property to see how it may impact customer journeys. Each event can be expanded by a single property. For example, users can break down the “add_to_cart” event by the “item_name” property to see the most common items added to cart.
To to do so, under path analysis click the “Expand Events” button, and choose the desired event:
Users can create Sankeys that appear in reverse order to focus on the end of the path and present the results from right to left. To enable reverse order, on the report settings turn on the “Reverse order” toggle:
Hide drop offs
For some of the journeys use cases, the focus is more on the journey’s side and less on the drop-offs. Now, with a simple click (under visualization settings) users can hide drop offs from the Path report visualizations.
The Export Segment enables exporting a set of users from CX Journeys platform to different destinations such as Google Bucket, S3, and SFTP (SSH).
Which formats are being supported? The supported formats are CSV and JSON.
Who can export segments? Each project user can export segments.
Which projects support import segment At the moment, we provide the solution for new project architecture. To check whether your project supports the export please contact Professional Services.
So how can I export a segment? Under the segments section, clicking on the 3 dots button (on the right corner for each segment) will open for you a drop down list, the 3rd option will be Export.
After clicking on export, the following screen appears:
Supported formats are CSV and JSON (with an option to Gzip the file)
In the export attributes option the users will have to pick the desired user attributes to be exported
Now you’re ready to move to the 2nd step!
In the 2nd step, the user will be able to choose one of the following options:
Destination URI – the location address of the bucket destination
Google bucket connection approach
CX Journeys Service Account Address – In the following approach users will
have to copy CX Journeys Address and paste it in the relevant bucket. In the bucket browser (in GCP) Under permissions tab, click add members, and pick a role for the permission. The minimum role needed is “Storage Legacy Bucket Owner”.
Another approach that can be used is creating a custom role (under roles section) with the following permissions:
After the role has been created, click add members, and pick the role created for the permission.
Service Account Key – In this approach, the user will have to create a service account in Google Cloud
Under IAM & Admin click Service Account, and then create service account
After completing step 1 and 2 (no need to add permission), In the last stage of the user will need to create the key in JSON format, and upload it to the wizard
The following fields should be filled:
Destination URI – The location address of the bucket destination
Region (default is US East N. Virginia)
Access Key ID – The user identifier for the bucket destination
Secret Access Key – The user password for the bucket destination
This option is fully customizable by the user, depending on the location of the address. The following fields should be filled:
File Path – The required folder for the SFTP file
Host Name – The location address of the SFTP server
Port (usually 22)
Once the configurations have been set, the user will be able Test the connection, and/or move to the last step.
The user will have to save two last preferences regarding the process:
Scheduling/one time operation (manual)
Notification to email whether the import failed for any reason
Save/Save and run now
The Import Segment enables importing a set of users to the CX Journeys platform, using manual file upload or the project Google Bucket source.
Which formats are being supported? The supported formats are CSV and JSON, however different extensions (json, jsn, csv, txt, gzip, zip & no extension) are being supported as well
Who can Import segments? Each project user can import segments
How can I import my own segment? Under the segments section, clicking on the + button (on the top right corner) will open you a drop down list, the 3rd option will be Import. User will be able to choose from two options
Import from Google Cloud Storage
Upload local file
Choosing the Google cloud storage shows the following screen:
File: clicking on the browse button will allow the user to choose the desired file from the account source bucket. If you don’t have access to the source bucket please reach your account admin or professional services.
Creating new user option allows the user to import new users to the platform
Update users option allows the user to update the users existing attributes
Unify users allows the user to unify multiple users records to a single user, according to the identity source logic. (Recommended state is No)
Manual upload approach holds the same functionality, the only difference is the size limit of 50MB.
After completing the 1st step, steps 2 are completely equal, between two options the user will reach to the mapping step.
On the left side the source file properties will be shown, while on the right side the tool will recommend the desired fields (properties) that will be mapped. In addition, for each property the platform will recommend the property expected type, which can be modified by the user (only for new properties).
For each property the user can choose the following options:
Add the property as a new property (whether there’s no existing property in the platform)
Ignore the property
Map the property to an existing property
Note: In the mapping stage the user has to map at least one identifier field (in the mapping fields)
Once the mapping have been set, the user will be able to reach the 3rd and final step.
The user will have need for last couple of preferences regarding the process:
Scheduling/one time operation (manual)
Notification to email whether the import failed for any reason
Save/Save and run now
Roles are a group of users who hold the same permissions, either for a project or dashboard. Once you add a role to the dashboard (or project), all the users of that role immediately get the same permissions.
Who can manage Roles? Roles are being managed by the Account Admin.
What is an Account? Account is the new parent hierarchy for project management. Means that one account can hold several projects which can be easily navigated between (Settings -> Active Project).
As a result, “Account Admin” user level has been added. For each account, only account admins will be able to create new roles.
How can I create a new role? Click on User icon -> Roles -> + Button
Each role can be related to several projects and/or several dashboards (left side). On the right side the account admin can add all the relevant users that will be part of the role.
Can I share a dashboard to roles? Sure, clicking on the dashboard share button will allow users to manage their share preferences for users and roles.
If I’m a project admin, can I provide a role permission? Yes. A role permission to a project can be provided (or removed) via the permissions section (Settings-> Permissions -> Roles Tab).
Improved KPI report functionality: added Having and Sort by capabilities
Google Standard SQL support (We continue to support Legacy SQL)
Date condition for merging Aggregation and Models tables
UI / UX improvements
Shipment and cache notifications
Ability to download models logs
New / Improved Features:
Path analysis visualization (Sankey) We are aware that the best way to get insights is by visualizing the data, so we’ve looked for a better way to present the users journey: Meet our new Sankey visualization!
It allows you to read the users actions in a simple and easy way and can help you maximize your conclusions about your users’ flow. It available in Path builder and CQL editor and the colors of the graph can be controlled through the visualization setting
New enhanced date picker We designed a new date picker that expands the querying abilities and improves the user experience.
now you can query your data with more time functions including:
– “Last N Days“: counting days back, not including the current day. In order to include the current day just set the “Include today” toggle ON.
– “Previous N Weeks / Months / Quarters / Years“: Calculates calendar time period, that do not include the current ones.
– “Current Week / Month / Quarter / Year“: calculates the current calendar time period.
– Custom date range by hours: adds the option to define start hour at the start date and end hour at the end date.
Additional customized chart color palettes We expanded our charts color palette and as from today, each report could be set with a different color palette to allow the best fit for your site’s design.
In order to choose a new palette, open the report and change the color palette on the visualization setting.
Improved KPI report functionality: added Having and Sort by capabilities Cooladata added an advanced menu to the KPI report which enables you to add “HAVING” and “SORT BY” SQL function to the KPI query.
The advanced menu will be shown at the bottom of the builder once the KPI report has at least one custom step and a breakdown.
This ability allows you to filter aggregated reports easily.
Firebase support Based on customer feedback and spotting an underserving in firebase analytics we released special features to help companies that rely on firebase as a development platform.
JSON flattening at the ETL level for deserializing nested JSON and matching it to the known schema of firebase. The result is the ability to query and not having to worry about unesting.
Multiple project consolidation – companies running two projects or more enjoy consolidation for enhanced querying and comparison. This is also a significant cost reduction.
Unlimited properties per event – 25 unique parameters with each event type. With Cooladata, there is no limit to the number of custom properties and dimensions per event.
Enriching from Firebase real-time database – we managed to combine and manage update all properties and dimensions from Firebase into the fact table. We make it more efficient, producing insights by merging both raw and real-time data.
For more information see our documentation.
Google Standard SQL support Up until now, Cooladata enabled running freehand queries over Cooladata using Standard SQL with CQL editors only. Cooladata now supports create a fully standard project, which automatically will support external tables with Standard dialect, support standard expressions and will allow querying the data with KPI report as part of a strategic plan to migrate Cooladata to operate in Standard SQL, encompassing all the advantages of the language, including performance enhancements, query optimizations and superior functionality.
For more information about the new Standard SQL functions (like “Unnest”, “With”, “Array” etc.) see our Standard SQL documentation.
Date condition for merging Aggregation and Models tables Up until now, using “Append and Update strategy” allowed you to update existing rows (according to a unique key) and append new rows to the destination table. Starting from toady, Cooladata allows adding to partitions tables (Aggregation Tables or Models) with “Append and Update strategy” days filter (based on the partitions column) which enables to replace only the latest tables’ rows. This feature improves executions’ performance and helps to manage huge tables.
UI / UX improvements: Saving lists filters during the session and open in a new tab We’ve listened to our customers’ requests and continue to improve our user interface. As parts of our efforts, we enabled to keep the list’s filters until the end of the session. This includes the search, column filter, sorting, the active tab and the number of the presented results in the page (page size). In addition, we allow opening reports, dashboards and tasks in a new tab using CTRL+Click/middle mouse click.
Shipment and cache notifications In order to improve query performance, Cooladata uses extensively caching and shipment.
Whenever a report is executed, Cooladata notify the user whether the data returned from the cache or used shipment
Controlling whether a report will use cache, as well as shipment, could have never been easier. Just click on the “Report Options” and set the toggle ON or OFF
Simplifying working with several projects in Cooladata
Cohort Auto Captions
Column filters for Table Visualization
ETL auto-population by scope ignoring NULL values
Lookup values for virtual properties and expressions
Standard SQL Up until now, Cooladata’s SQL was compatible with Google BigQuery SQL Legacy dialect. Cooladata now enables running freehand queries over Cooladata using Standard SQL, as part of a strategic plan to migrate Cooladata to operate solely in Standard SQL, encompassing all the advantages of the language, including performance enhancements, query optimizations and superior functionality. Writing Standard SQL is now available when creating a CQL report, aggregation table, model, CQL segment or alert. For more information see our Standard SQL documentation.
Partitioning in External Tables Cooladata’s events and sessions tables are automatically partitioned by time. From now on, Cooladata also enables the partitioning of other external tables in your project by a timestamp or date column. Partitioning makes sure you run only on the selected time range in the table instead of the entire table, improving performance and minimizing query run time. Partitioning is available only when creating aggregation tables or models using Standard SQL dialect, and are only available for querying in Standard SQL dialect.
Merge data using Integrations Integrations fetching external data can now write to a table in “Append and Update strategy”, which will update existing rows (according to a unique key) and append new rows to the destination table. This can reduce the number of rows fetched and integrated in each run to only the updated and new rows.
Writing Models to your external destination Models can now save their results to external destinations as available in aggregation tables: your Google Bucket or BigQuery dataset.
Simplifying working with several projects in Cooladata We’ve listened to our customers’ requests and enabled filtering all the dashboards and reports shown on the lists filtered by the active project. Since you can add reports from several projects to the same dashboard, some dashboards might be shown in a few projects. When creating filters, make sure you choose the right project and table to filter by in the filter widget.
Chart Visualization – Choose whether or not to auto-fit your chart We are aware that sometimes, it’s the big picture showing big trends that is important in a chart, whereas in other times – it’s the little details we want our data consumers to focus on. That’s exactly why we have enabled auto-fitting a single chart in the report size – so you can choose how your report will be viewed in a dashboard.
Cohort Auto Captions The cohort widget now auto-populates the cohort captions that explain to your report consumers what it is they are seeing. You can override the captions manually as you wish.
ETL auto-population by scope ignoring NULL values Sending an explicit Null for a session or user scope property will be ignored and the last non-null value sent will be auto-populated (“smeared”) going forward.
Column filters for Table Visualization Filtering a single value has never been easier. Just click on a single column header and choose the value you want to see.
Lookup values for virtual properties and expressions Lookup values are the list of values in the filter drop down. These are populated as part of the ETL, so they weren’t available for virtual properties and expressions until now. From now on, all properties can have a drop down list of values in filters.