Term | Description |
Account admin | Account admin is a role given to the user who creates and sets up the DataPrep organization for your organization. An organization is usually tied to one organization |
Add data | The Add data button is available on the Visual pipeline builder. You can click this button to import data. |
Base datatype | These are the fundamental data types upon which a few other data types are built. Example: Text |
Base unit | In DataPrep, conversions between different units and the histogram representation of data distribution is applied on top of the base unit. For example, for the Temperature datatype, the base unit is celcius (°C). |
Basic Auth | Basic Auth is an authorization type that requires a 'Username' and 'Password' to import data from URL |
Bucketing | The bucketing technique groups values in multiple ranges or buckets for easier understanding of highly scaled data. The granular details of such data can prevent us from getting a broad view of the data. This form of grouping values in a range not only helps in observing the overall view of the data, but also in finding data patterns. |
Column header | Column header is the name of the columns in a dataset. |
Column level transform | These are the transforms that can be applied on one or more columns and not on the whole data. Example: Change data type |
Configure schedule | When you export a pipeline, you can repeat it by creating a schedule using the Schedule option. You can select the required Repeat method and the time to repeat and create an export schedule. |
The context menu can be seen when you right-click the column name in a dataset. This menu provides various column level transforms along with a few icons to sort, delete, rename and move columns. | |
Count | The count transform allows you to count the occurrence of data in a column based on the condition given. You can count the values that match your selection based on the different patterns that fit your selection. |
Criteria expression | When you apply multiple conditions while applying a transform on a dataset, you can combine them using the required logical operators. The final expression of the conditions is displayed under the Criteria expression field. |
Custom data type | Custom data types are used to validate organization-specific data, such as employee ID, invoice ID, shipment tracking ID, or asset ID. By creating a custom data type, you can set a standard for your organization-specific columns in your data. |
Custom date format | The custom date format creator consists of the date and date-time tiles required to construct a date format. You can simply click the required elements to form or customize a date or a date-time format instead of manually writing the format yourselves.This appears when you choose to customize a date format. |
Custom range | Custom range is an option available in the bucketing technique in which you can input your own conditions to determine bucket labels when you group values. |
Data catalog | Data catalog helps in controlling the usage of datasets, their status, and discovery. This includes scheduling and sharing of datasets depending on their readiness for use. This section is visible to only the account admin and the organization admins. |
Data cataloging | Data cataloging in Zoho DataPrep helps with data management and data discovery depending on the usage of data assets, their status, and associated information. It can be used to evaluate and implement the readiness of data for use, by intended users in an organization. |
A data quality bar represents the quality of data in each column. It splits data quality into valid data, invalid data, and missing values, based on data type of the column. Green represents valid data, red represents invalid data, and grey represents the missing values. | |
Zoho DataPrep allows you to trigger your pipeline using Webhooks. A webhook is a user-defined HTTP callback that is triggered when a particular event occurs at the source site. | |
The data source configuration icon is displayed in the Ruleset pane. This will have details such as the source of the input data which includes source type, connection details, timestamps of last import and next import, parsing details, and if the import is scheduled. You can also edit your import schedules from here. | |
Reload data | You can manually refresh your stage with the latest data by reloading data from your data source. You can use the manual reload option to fetch data that was failed during a schedule. |
Data type is a key attribute associated with a column using which the valid and invalid data are classified and data quality is calculated. | |
These are the transforms that can be applied on the dataset as a whole. Example: Join | |
Datasets | A dataset is a collection of discrete items of related data. |
Date picker | Date picker is a collection of format strings of date-time components in the form of tiles to choose from. All the supported format strings are incorporated into the date picker. You can simply click to select the required element to form or customize a date or a date-time format instead of manually writing the format string yourselves. |
Delimiter | A delimiter is a separator, usually a symbol or a whitespace that separates one date time element from the other. For example, in the date: 11/24/2020 , / is the delimiter that separates the date elements. |
Derived datatype | These are the data types derived from the base data types. For example, currency is a data type that is derived from the decimal data type. |
Format string | Format strings are the date and date-time elements such as yy,dd,mm. A date format is formed using various format strings. |
Intelligent suggestions are the transforms suggested by DataPrep when one or more columns are selected. The suggestions are shown based on the imported data for effective data preparation. | |
Join potential | When you perform a join transform, DataPrep automatically calculates and shows the Join potential of both datasets in percentage making it easier for you to choose a dataset for the Join transform. |
Mark as default | You can mark a workspace as default using this option in the workspace details page and DataPrep will automatically open this workspace for you when you launch. |
Backfill | Backfill run can process data that was missed in previous schedules due to a change in data models or data preparation workflow. |
Pipeline builder is a place that allows the users to create data flows in DataPrep, with multiple data stages and various flow level transforms. | |
Match accuracy | The match accuracy is shown when you try to change a data type of a column. Match accuracy is the percentage of data volume in the selected column that conforms to a data type. |
No Auth | When you import data from URL that does not require any authorization you can use this option in the Data Import page. |
OAuth2.0 | OAuth2.0 is an authentication technique used when you import data from a URL. It requires certain details such as ClientID, ClientSecret, Access Token, and Refresh token to authorize the URL to allow import. |
Organization | An Organization is set up by the Account admin in DataPrep. An organization is tied to one account and the Account admin will have full access to perform read, write and share operations in all workspaces, datasets, ruleset templates, and other entities such as the saved connections and user groups of their organization. |
Organization admins are designated by the Account admin during the DataPrep setup process. An organization admin is a user responsible for creating and managing the workspaces in Zoho DataPrep. | |
A pipeline execution is called a job, a job tracks the progress of imports, transformations and exports in a pipeline among other stats. | |
Re-authenticate | The Re-authenticate link appears on the schedule that has failed either because the schedule creator is deactivated or removed from the organization, or the connection used is deleted or no longer shared with the user. |
Repeat method | Repeat method is the time taken for one schedule. For example, when you want the import to be done every 24 hours, you can choose 'Every N Hours' and '24 Hours' options when you create a schedule. |
Each transformation applied on the dataset is stored in DataPrep as a rule, in order of their execution. The ordered list of these rules is called a Ruleset. | |
The Ruleset history includes user information such as, who created or modified the rules, the timestamp when the changes were made, and the various operations added to the Ruleset. | |
Ruleset pane | The data source and the rules applied to a dataset are displayed in the Ruleset pane. You can click the Ruleset icon in the top-right corner of the Studio page to view the Ruleset pane. |
Ruleset templates are the Rulesets from various datasets saved for reuse. | |
The connections that you create are saved in DataPrep are saved for using them later. | |
The Schedule option is used to schedule the import from various sources, apply the transforms, and export to the destinations in your pipeline at regular intervals. | |
Pipeline | Pipelines are fundamental entities that allows users to create data flows in DataPrep, with multiple data stages and various flow level transforms. |
Smart selection is a feature in DataPrep that offers you an array of suggestions using the pattern matching notations when you select portions of the column data that you wish to transform. | |
The studio page appears when you open a dataset. The studio page is the place where you can apply transforms to a dataset. | |
Studio panel | The Studio panel appears on the right side of the Studio page and this is the place where you apply transforms. |
Manual run | A manual run will run your pipeline on the existing data without refreshing, which means no new data will be imported. |
Tags | You can associate tags to entities such as workspaces and datasets to identify, classify, or group them. You can also use tags as labels to quickly search for your entities in DataPrep. |
Zoho Flow | Zoho DataPrep allows you to trigger your pipeline using Zoho Flow. Zoho Flow is an integration platform which connects cloud applications. It helps you set up workflows to automate information exchange among the apps you use. |
Target matching | Target matching allows you to set a target and align the source dataset to match with your target before exporting data. You can import a target dataset to match the column structure, formats, and data types in the existing source dataset. This helps you ensure seamless exporting of high-quality data to your destination. |
The topbar appears on top of the Studio page and it has the dataset switcher and menu bar on the right. | |
Truncate | You can use the Truncate operation to remove unnecessary characters or words from the end of text values in your data. |
Users can create user groups to share entities. User groups can be created to share entities such as workspaces, datasets, ruleset templates and saved data connections to a group of users. | |
Wildcard tab | You can use the wildcard tab if you want to apply conditions and construct your own expression using the logical operators. |
Workspace | Workspaces are entities that hold the various datasets imported into Zoho DataPrep. |
Workspace shared user | Zoho DataPrep has roles for users based on their entity permissions. When a workspace is shared with a user, they will have certain permissions and restrictions to perform entity-related activities in DataPrep. |
The workspace details page lists all relevant information about the workspace and its child entities. It includes details on the datasets, ruleset templates, and the sync status of each dataset within the workspace. | |
Zia is a search functionality using which you can search for any entities across Zoho DataPrep. | |
Zoho Analytics | Zoho Analytics is a self-service BI and data analytics software that lets you visually analyze your data, create stunning data visualizations and discover hidden insights from your data. |
Zoho Databridge is a lightweight independent utility that connects your on-premise data source and Zoho applications to import data. In other words, you can quickly and securely transfer data from your internal network to Zoho applications. | |
Zoho Workdrive | Zoho Workdrive is an online file synchronization, storage, and content collaboration platform. You can use this cloud storage service to import data. |
Zoho CRM | Customer Relationship Management (CRM) is kind of a software platform that businesses can use as a repository to unite and streamline sales, marketing, and customer support activities. It's the go-to solution for businesses to store their customer data while keeping track of their customer interactions. |
Learn how to use the best tools for sales force automation and better customer engagement from Zoho's implementation specialists.
If you'd like a personalized walk-through of our data preparation tool, please request a demo and we'll be happy to show you how to get the best out of Zoho DataPrep.
You are currently viewing the help pages of Qntrl’s earlier version. Click here to view our latest version—Qntrl 3.0's help articles.