RapidCanvas Docs
  • Welcome
  • GETTING STARTED
    • Quick start guide
    • Introduction to RapidCanvas
    • RapidCanvas Concepts
    • Accessing the platform
  • BASIC
    • Projects
      • Projects Overview
        • Creating a project
        • Reviewing the Projects listing page
        • Duplicating a Project
        • Modifying the project settings
        • Deleting Project(s)
        • Configuring global variables at the project level
        • Working on a project
        • Generating the about content for the project
        • Generating AI snippets for each node on the Canvas
        • Marking & Unmarking a Project as Favorite
      • Canvas overview
        • Shortcut options on canvas
        • Queuing the Recipes
        • Bulk Deletion of Canvas Nodes
        • AI Guide
      • Recipes
        • AI-assisted recipe
        • Rapid model recipe
        • Template recipe
        • Code Recipe
        • RAG Recipes
      • Scheduler overview
        • Creating a scheduler
        • Running the scheduler manually
        • Managing schedulers in a project
        • Viewing the schedulers in a project
        • Viewing the run history of a specific scheduler
        • Publishing the updated data pipeline to selected jobs from canvas
        • Fetching the latest data pipeline to a specific scheduler
        • Comparing the canvas of the scheduler with current canvas of the project
      • Predictions
        • Manual Prediction
        • Prediction Scheduler
      • Segments and Scenarios
      • DataApps
        • Model DataApp
        • Project Canvas Datasets
        • Custom Uploaded Datasets
        • SQL Sources
        • Documents and PDFs
        • Prediction Service
        • Scheduler
        • Import DataApp
    • Connectors
      • Importing dataset(s) from the local system
      • Importing Text Files from the Local System
      • Connectors overview
      • Connect to external connectors
        • Importing data from Google Cloud Storage (GCS)
        • Importing data from Amazon S3
        • Importing data from Azure Blob
        • Importing data from Mongo DB
        • Importing data from Snowflake
        • Importing data from MySQL
        • Importing data from Amazon Redshift
        • Importing data from Fivetran connectors
    • Workspaces
      • User roles and permissions
    • Artifacts & Models
      • Adding Artifacts at the Project Level
      • Adding Models at the Project Level
      • Creating an artifact at the workspace level
      • Managing artifacts at the workspace level
      • Managing Models at the Workspace Level
      • Prediction services
    • Environments Overview
      • Creating an environment
      • Editing the environment details
      • Deleting an environment
      • Monitoring the resource utilization in an environment
  • ADVANCED
    • Starter Guide
      • Quick Start
    • Setup and Installation
      • Installing and setting up the SDK
    • Helper Functions
    • Notebook Guide
      • Introduction
      • Create a template
      • Code Snippets
      • DataApps
      • Prediction Service
      • How to
        • How to Authenticate
        • Create a new project
        • Create a Custom Environment
        • Add a dataset
        • Add a recipe to the dataset
        • Manage cloud connection
        • Code recipes
        • Display a template on the UI
        • Create Global Variables
        • Scheduler
        • Create new scenarios
        • Create Template
        • Use a template in a flow notebook
      • Reference Implementations
        • DataApps
        • Artifacts
        • Connectors
        • Feature Store
        • ML model
        • ML Pipeline
        • Multiple Files
      • Sample Projects
        • Model build and predict
    • Rapid Rag
  • Additional Reading
    • Release Notes
      • May 14, 2025
      • April 21, 2025
      • April 01, 2025
      • Mar 18, 2025
      • Feb 27, 2025
      • Jan 27, 2025
      • Dec 26, 2024
      • Nov 26, 2024
      • Oct 24, 2024
      • Sep 11, 2024
        • Aug 08, 2024
      • Aug 29, 2024
      • July 18, 2024
      • July 03, 2024
      • June 19, 2024
      • May 30, 2024
      • May 15, 2024
      • April 17, 2024
      • Mar 28, 2024
      • Mar 20, 2024
      • Feb 28, 2024
      • Feb 19, 2024
      • Jan 30, 2024
      • Jan 16, 2024
      • Dec 12, 2023
      • Nov 07, 2023
      • Oct 25, 2023
      • Oct 01, 2024
    • Glossary
Powered by GitBook
On this page
  • Viewing options in the side panel of a dataset block
  • Viewing the dataset information
  • Exporting a dataset to the local system
  • Deleting the uploaded dataset
  • Reloading latest datasets from fivetran connector
  • Configuring Destination Details for Output Datasets
  • Steps to Configure Destination Details:
  • Additional Actions:
  1. BASIC
  2. Connectors

Connect to external connectors

PreviousConnectors overviewNextImporting data from Google Cloud Storage (GCS)

Last updated 1 month ago

RapidCanvas connectors module enables you to interact with different external connectors to import data into the platform and make predictions on this data with the built machine learning models.

List of connectors supported:

  • Google cloud storage. For more information, see .

  • Amazon S3. For more information, see .

  • Azure blob. For more information, see .

  • MongoDB. For more information, see .

  • Snowflake. For more information, see.

  • MySQL. For more information, see .

  • Amazon Redshift. For more information, see .

  • Fivetran connectors. Example : Google Drive. For more information, see .

Viewing options in the side panel of a dataset block

Use this procedure to view all actions you can perform through pull-out window of a dataset.

To view dataset options:

  1. Click on the dataset block on the canvas. This opens the pull-out window.

  2. Perform any of these actions:

  • Click Preview to view the data in the file you have uploaded. For more information, see .

  • Click the AI Guide icon to navigate to the AI Guide to query about the dataset.

  • Click the plus icon to select Template Recipe, AI-assisted Recipe, Code, or Rapid Model Recipe options. For more information, see .

  • View the significance of the dataset node. This AI snippet is only visible when the AI snippet generator is run. You can always update the content using the Update content option.

  • Click the Actions button and select Export to download the file to your local system in CSV format. Note that you can only export a file as CSV if it is a source file. If you need to export the output dataset generated after running the recipe, you can choose between two formats: Parquet or CSV. For more information, see .

  1. Review these details:

  • Created: The date on which the file was uploaded.

  • Updated: The date on which the file was last updated.

  • Total size: The total file size.

  • Rows: The total number of rows in the file.

  • Columns: The total number of columns in the file.

  • Source: The source from where the dataset has been imported. Clicking on the file import will redirect you to the View Data page.

  1. View the dataset summary in the Summary section. This section provides key details about the dataset, including:

    • Column – Names of the columns in the dataset.

    • Data Type – The data type of each column.

    • Null Percentage – The percentage of missing values in each column.

    • Unique Values – The number of unique values in each column.

    • First Row – The values in the first row of the dataset.

    • Last Row – The values in the last row of the dataset.

If the dataset has been imported from connectors, you can view the following source details:

  • Data connector name: The name of the connector. Clicking on the link will redirect you to the data connector configuration page.

  • Type: The type of connector.

If this is an output dataset, you can view the Destination Details section:

  • Select the data connector to which you want to send the output dataset.

    • If the selected connector is GCS, specify the destination folder name and destination file name, then click Save to save these details. Later, you can export the file. You can also use the Delete option to remove the configured destination.

    • If the selected connector is MySQL, specify the table name. There are two options:

      • Append: Appends the dataset to the existing table, provided both have the same schema.

      • Replace: Replaces the existing data and schema with the new one.

      • Once configured, click Save to save the details. Later, you can export the file.

    • If the selected connector is Snowflake, specify the table name, database name, schema, warehouse, and role. There are two options:

      • Append: Appends the dataset to the existing table, provided both have the same schema.

      • Replace: Replaces the existing data and schema with the new one.

      • Once configured, click Save to save the details. Later, you can export the file.

Note: Fivetran connectors cannot be configured in the Destination Details section.

Viewing the dataset information

Use this procedure to view dataset details and perform various actions.

To view data:

  1. Select the dataset block uploaded onto the canvas. This opens the pull-out window.

  2. Click Preview to navigate to the Data page. The dataset records are displayed in a tabular format.

  3. View the data in each column under the Data tab:

    • Check the data type associated with each column.

    • Click the Data Summary button to view a detailed summary of each column in the dataset. This includes:

      • Unique Value Count – The number of distinct values in a selected column.

      • Pie Chart for Boolean Values – A visual representation of Boolean (true/false) values.

      • Top Two Categories – The two most frequent categories in a column, along with their counts.

      • Histogram for Numeric Data – A graphical distribution of numerical values.

      • Null Percentage – The percentage of missing values in each column.

    • Use the search box to find specific terms and clear the search when needed.

Column Options

Click the ellipses icon next to each column to:

  • Hide – Hide the column in the table.

  • Sort ascending – Sort the column in ascending order.

  • Sort descending – Sort the column in descending order.

  • Search column – Search for a specific column name.

Additional options:

  • Click Resize Columns to adjust column width.

  • Click Schema to view the data type of each column in the dataset.

  • Use the search function to type a keyword and retrieve relevant data from the table.

  1. Under the Data Analysis tab, analyse the dataset to identify:

  • Missing values

  • Total variables (numeric, text, categorical)

  • Total observations

  • Duplicate values

You can also generate charts to gain insights from the complex data. The charts also help you understand the data patterns and trends required to perform feature engineering. You can generate four charts initially, and doing advanced analysis gives you up to 10 charts by sending the dataset to AI.

  1. Navigate to the Correlation tab to extract correlations and relationships within the data. The correlation heat map shows how each variable in the dataset is correlated with one another, as a color-coded matrix.

  2. Under the Alerts sub-tab in Data Analysis, view alerts and associated tags.

  3. Click the drop-down to view the Segments and Source tabs.

  4. Click the Source tab. Please note that this tab is only visible for SQL based connectors. This allows you to modify the SQL query directly within the interface, ensuring that your dataset reflects the latest query data. Any changes made to the query will automatically update the dataset with the most current data available. However, updating the query data on the data pipeline will cause all related recipes linked to this dataset to move to an unbuilt state. To incorporate the changes, you will need to re-run the entire data flow.

  5. Click Save Query to update the query. Doing this replaces the existing data with the latest query data.

  6. Click Yes, Replace on the dialog box. This updates the existing query with new SQL query.

  7. Click Audit History in the Actions tab to check the log of user activities. Each entry will include the user who performed the action, type of action performed and timestamp.

  8. Review the source details:

Field

Description

Source Type

SQL-based connector type.

Connection Name

Name of the connector.

Updated on

Date the connector was last updated.

Query

Click Edit to modify the SQL query and run it.

Important: Updating the query in the data pipeline moves all related recipes to an unbuilt state, requiring a re-run of the data flow.

  1. Select Segments from the drop-down to open the Segments tab and view the associated segments:

Field
Description

Description

The description of the segment.

Name

The name of the segment.

Created

The date and time the segment was created.

Rows

The row limit for segmentation.

Actions

Edit or delete the segment.

You can perform the following actions on the data view page clicking the plus icon:

  • Add a template recipe: Use the Template option.

  • Use AI-Assisted recipe: Use the AI-assisted option.

  • Run Rapid Model recipe: Use the Rapid Model option.

  • Add a segment: Available only for the source dataset.

You can perform the following action clicking the Actions drop-down:

  • Download the dataset, using the Export option.

  • Use the Export Search Results option in the Data tab to export only the results for your specific search term. After conducting a search, you can export just the results obtained for that term

  • Delete the dataset and associated recipes with the dataset, using the Delete option.

You can also use the AI guide to query on this dataset, using the AI guide icon.

Exporting a dataset to the local system

Use this procedure to download the input and output dataset to your local system in the csv file format.

To export a dataset:

  1. Select the dataset block that you have uploaded onto the canvas. This opens the pull-out window.

  2. Click Export to download the dataset in csv format to your local system.

You can also export the dataset from the dataset page, using the EXPORT option. This dataset page is displayed clicking VIEW DATA in the pull-out window.

Deleting the uploaded dataset

Use this procedure to delete a dataset block from the canvas.

To delete a dataset block:

  1. Select the dataset block that you want to delete from the canvas. This opens the pull-out window.

  2. Click Delete to delete the dataset.

  3. A dialog box prompts that deleting the dataset also deletes the recipes associated with it.

  4. Click Delete to delete the dataset permanently from the project or click Cancel to discard the action.

Reloading latest datasets from fivetran connector

Use this procedure to reload fresh data from the fivetran connector. This dataset syncs with the remote storage and retrieves the latest dataset.

To reload a dataset:

  1. Select the dataset block that you want to reload from the canvas. This opens the pull-out window.

  2. Click Reload to reload the dataset.

  3. A dialog appears. Click Reload to fetch the latest dataset. Ensure that the schema of this dataset same as the current one.

Configuring Destination Details for Output Datasets

When working on the Canvas, you may generate an output dataset. You can configure the Destination Details to specify where the latest output dataset will be stored each time the canvas flow runs.

Steps to Configure Destination Details:

  1. Select the Output Dataset – Click on the output dataset block on the canvas to open the side sheet.

  2. Choose a Data Connector – Select the connector where you want to save the output dataset. The available fields vary based on the selected connector.

  3. Specify Table Name (for MySQL Connector) – If using a MySQL database connector, enter the table name where the output dataset should be stored.

  4. Set Save Mode – Choose either:

    • Append – Adds new data to the existing dataset, provided both share the same schema.

    • Replace – Replaces the existing dataset with the latest output.

  5. Save the Configuration – Click Save to apply the changes. You can then use the Export option to send the dataset to the selected connector.

Once saved, the destination connector appears as a node on the canvas, serving as a visual indicator that the output dataset is linked to a destination.

Additional Actions:

  • Delete Connector – Remove the connector linked to the output dataset using the Delete button.

  • Preview Data – Click Preview to view the contents of the dataset.

  • Use AI Guide – Get AI-generated prompts to explore and analyze the dataset.

  • Run Recipes – Click the + (plus) button to apply different recipes and perform data transformations.

  • Export Dataset – Click the ellipses (â‹®) icon to export the dataset as CSV or Parquet.

  • Delete Dataset – Click the ellipses (â‹®) icon to remove the output dataset from the canvas.

Add: Add a new file to the dataset (default is Append). See .

Appending a file to the source dataset
Import data from Google Cloud Storage
Import data from Amazon S3
Import data from Azure Blob
Import data from Mongo DB
Import data from Snowflake
Import data from MySQL
Import data from Amazon Redshift
Import data from Google Drive
Recipes
Viewing the dataset information
Exporting a dataset to the local system
Export Dataset
Delete Dataset
Reload Dataset