Process Types
📄️ AA Loader
The Adobe Analytics or AA Loader process provides the ability to bring raw Adobe Analytics clickstream data into a customer environment from the specified location where Adobe places the files (i.e. S3 bucket or SFTP folder). Once the hit\_data and lookup files are placed into a specified location on the environment, the data can be used by the Syntasa Event Enrich process, and external tables built providing ability to query.
📄️ Adobe API
The Adobe API process provides the ability to pull in the Adobe custom variable definitions into the Adobe Analytics Input Adapter app creating a Report Schema output that provides the details and rules for each eVar, Prop, and Event. Custom variable definitions allow the ability to create user-friendly labels in the Syntasa processed data. Connecting the Adobe API also allows for Syntasa to run queries against the Adobe Analytics reporting API and turn on the automated audit comparing Adobe and Syntasa data. Report Schema process output node connects directly to the Event Enrich node.
📄️ Adobe Audit
Adobe Audit is one of the many processes available in the Syntasa platform that users can utilize while creating apps. The process is available in all three modules, Synthesizer, Composer, and Orchestrator.
📄️ Adobe Classifications Export
The Adobe Classification Export process allows us to request a classification report from Adobe via API calls. To utilize this process, the Adobe API connection needs to be configured. Then we can make use of the connection by using the Adobe API process.
📄️ Adobe Event Enrich
Using data prepared by the AA loader process, the Event Enrich process applies functions to the data, joins lookups, and writes the data into an event-level dataset. The dataset is the foundation for building the session, product, and visitor datasets, but can also be used for analysis and user-defined analytics datasets.
📄️ Adobe Product Enrich
Using data prepared by the Event Enrich process, the Adobe Product Enrich process applies functions to the data, joins lookups and writes the data into a product level dataset, which can be thought of as a sub-event level.
📄️ Alert
The purpose of the Alert process is to provide a mechanism for evaluating a variable or set of variables against a threshold and provide an email or Slack message when that threshold is met. For example, if a model is running and the analysts needs to know if the grade of the model gets to a certain so they can re-tune it, then the Alert would provide this so the analyst does not have to go in and manually monitor.
📄️ BQ Process
The BigQuery (BQ) Process is a Synthesizer, Composer, and Orchestrator process that builds and manages user-defined analytics datasets using pre-written and verified working code. It is designed to allow the user to query an existing BigQuery table to create a new dataset within the Syntasa environment.
📄️ Custom Shell
Description
📄️ Dash
With a focus on visuals, the Dash Process enables data scientists to create dashboards from the data processed through their pipeline rather than just dataset tables. This is an additional tool that highlights data and turning this data into dashboards, adding to the capability of dashboards.
📄️ Decision Tree
Description
📄️ Event Loader
The purpose of the Event Loader process is to ingest non-Adobe raw files, such as marketing data or enterprise lookup data.
📄️ Evaluator
Description
📄️ Feature Learn & Feature Transform
When dealing with different data science use cases, we come across different kinds of variables such as strings, numbers, or text data. Datasets usually have a mix of categorical and continuous variables. However, algorithms that are powered by these datasets understand only numbers. Hence, it is important to convert and transform all the variables into numbers which can then be fed into an algorithmic model. Additionally, these transformations on source data also act as foundation datasets for feature engineering which one of the most fundamental aspects of machine learning.
📄️ Featurize
Description
📄️ From BQ
The From BQ provides the ability to load data from a BigQuery table into a Hive structured storage location in the Syntasa environment. Sample use cases may include:
📄️ From DB
The purpose of the From DB process is to define the input database table, schema, and load into a big data environment. A database connection will need to be defined first on the Connections page and that Database Connection will need to be dragged onto the canvas and connected to the From DB process.
📄️ GA Output Adapter
The GA Adaptor provides the ability to configure and import a file to the Google Analytics Data Import. Data Import provides a means to upload external data (i.e. CRM, product catalog, model scores) to GA providing metadata that helps to add context to the Analytics behavioral data.
📄️ GA Report
A process to download a customized report from Google Analytics.
📄️ Generalized Linear Regression
Description
📄️ Generic Event Enrich
The Generic Event Enrich process applies functions to the data, joins lookups and writes the data into an event level dataset. The dataset is the foundation for building the session, product, and visitor datasets, but can also be used for analysis and user-defined analytics datasets.
📄️ Gradient Boosting Tree
Description
📄️ LTM
Description
📄️ Lookahead
The Lookahead process creates the label dataset which contains the outcomes to predict. A label is a defined success event that the model learns from. It is possible to have more than one label for modeling.
📄️ Lookback
The Lookback process is used to build a historical dataset for applications like the propensity scoring app. This dataset includes all the features needed to train a machine learning model by looking back a specified number of days from the processing time and creating a table with user-defined columns and filters.
📄️ Logistic Regression
Description
📄️ ML Engine Prediction
Description
📄️ ML Engine Train
Description
📄️ Matomo Loader
The purpose of the Matomo Loader process is to get data from the Matomo website
📄️ Orchestrator Free Form
"Orchestrator Apps" refer to applications that create integrations with enterprise systems, enabling them to deliver data points. Orchestrator Free Form is an Orchestrator app that provides a blank workflow for the ability to create custom apps by using pre-built processes like To DB, Post, and From File.
📄️ Post
The Post is an Orchestrator process that facilitates the building of a Customer Intelligence Hub by pushing data out to non-relational databases such as Big Table and HBase. Some of the high-level capabilities available with the Post are:
📄️ Quantile Evaluator
The Quantile Evaluator on the Syntasa platform calculates cumulative performance metrics based on score percentiles.
📄️ Quantile Evaluator by Partition
The Quantile Evaluator on the Syntasa platform calculates cumulative performance metrics by score percentile and date.
📄️ Random Forest
Description
📄️ Registry Identity
The Register Identity process provides the user the ability to identify and register identity pairs with the Syntasa Identity Graph. An identity graph's contents include the identity pair, first time that pair was observed and whether that pair is suspect or not. The identity pair consists of a local\_id
📄️ Score
Description
📄️ Session Enrich
The Session Enrich process applies functions to data generated by the Event Enrich process, joins lookups, and writes the data into a session-level dataset, which can be thought of as event data aggregated at the visit level. This data is then grouped by day to allow for partitioning, which allows for more rapid and easy analysis over specified time periods.
📄️ Spark Learn
Description
📄️ Spark Score
Description
📄️ Split
Description
📄️ To DB
This process provides the ability to write data to a downstream database.
📄️ To File (aka Publish)
To File is an Orchestrator process that facilitates pushing data out to external systems and storage locations. Some of the high-level capabilities available with the To File process are:
📄️ Transform
The Transform is a Synthesizer, Composer, and Orchestrator process that helps to build user-defined analytics datasets. Some of the high-level capabilities available with Transform are:
📄️ Unified Event Enrich
The Unified Event Enrich process takes the data in raw and enriches and transforms it to follow the SYNTASA schema as defined in the configuration mapping unique to each data source.
📄️ Unified Product Enrich
The Unified Product Enrich process uses data prepared by the Unified Event Enrich process to apply functions to the data, connect lookups, and publish the data into a product-level dataset, which can be thought of as a sub-event level.
📄️ Unified Session Enrich
The Unified Session Enrich process applies functions to data generated by the Event Enrich process, joins lookups, and writes the data into a session-level dataset, which can be thought of as event data aggregated at the visit level. This data is then grouped by day to allow for partitioning, which allows for more rapid and easy analysis over specified time periods.
📄️ Unified Visitor Enrich
Using data prepared by the Unified Event Enrich process, the Unified Visitor Enrich process applies functions to the data, joins lookups, and writes the data into a visitor-level dataset, which can be thought of as event data aggregated at the visitor level.
📄️ Using Code Processes
Pre-requisites}
📄️ Visitor Enrich
Using data prepared by the Event Enrich process, the Visitor Enrich process applies functions to the data, joins lookups and writes the data into a visitor level dataset, which can be thought of as a event data aggregated at the visitor level. This data is then grouped by day to provide ability to partition, which provides the ability to more quickly and easily analyze across specific time periods. As the user, you will find the data configured where there is one record per visitor ID per day the visitor ID was present.