Team plan

Team plan is a paid service offered on KNIME Community Hub.

This allows you to create a team on KNIME Hub to collaborate on a project with your colleagues. Additionally, you can add execution capabilities to your team. This will allow the team members to schedule and automate the execution of workflows on KNIME Community Hub.

A team is a group of users that work together on shared projects. Specific Hub resources can be owned by a team (e.g., spaces and the contained workflows, files, or components) so that the team members will have access to these resources.

You can purchase a team and choose the size of the team, meaning the number of users that will be able to access the team resources, and the disk space that you need to save your team workflows, data, and components.

The team can own public or private spaces. For more details, see the section Team owned spaces.

The items that are stored in a team’s public space will be accessible to everyone and be presented as search results when searching on KNIME Community Hub. However, only team members will have upload rights to the public spaces of the team.

Also, since only the team members have read access to the items that are stored in a team’s private space, KNIME Community Hub, users that are part of a team can collaborate privately on a project.

You can also add to your team the ability to run workflows and also to schedule the execution of the workflows at specific intervals, easily from the web user interface of KNIME Community Hub.

Create a team

To create a team, you need to subscribe to a Team plan. To do so, sign in with your KNIME Hub account and navigate to Pricing in the top right corner of the page, or go to KNIME Hub pricing page. Here you can proceed with the subscription to the Team plan.

To subscribe to a Team plan, you need to have a KNIME user account.

You will be asked to provide the details of the account and your payment information. Then, you can choose your subscription plan. You can choose how much disk storage and how many users you would like to purchase. The basic plan consists of 30GB of disk storage and a total of 3 users. Please note that the subscription will be automatically renewed every month if not canceled, and that users and disk storage can be adjusted later. Your usage will be prorated for the next billing cycle. To change or cancel your subscription, follow the instructions in the section Manage team subscription.

Once you have successfully purchased a team, you can assign it a name and start adding members to your team. The number of members that can be added to the team is limited to the number of users that you purchased.

There are two types of roles a user can be assigned when part of a team:

The team creator is automatically assigned the administrator role and can promote any of the team members to administrators. In order to do so, please follow the instructions in the section Manage team members.

Team owned spaces

A team can own an unlimited number of both public and private spaces.

  • Team owned public spaces: The items that are stored in a team’s public space will be accessible by everyone and be presented as search results when searching on KNIME Community Hub. Only team members will have upload rights to the public spaces of the team.

  • Team owned private spaces: Only the team members have read access to the items that are stored in a team’s private space. This will then allow KNIME Community Hub users that are part of a team to collaborate privately on a project.

You can create a new space by going to the team’s profile. To do so, click your profile icon in the top right corner of KNIME Hub and select your team. In the tile Create new space, click Private space to create a private space for your team, or Public space to create a public space. You can then change the name of the space, or add a description. You can change or add these also later by going to the relative space page and clicking the space name or Add description button to add a description for the space.

Furthermore, you can change the visibility of the space from private to public and vice versa, or delete the space. To do so, from the space page, click the 16 icon, as shown in the image below.

img change visibility delete space
Figure 1. Change visibility or delete an existing team owned space

Manage space access

You can also manage the access to a specific space. To do so, navigate to the space and click the 16 icon.

img manage space access pencil
Figure 2. Open the Manage space access side panel

In the Manage space access side panel that opens, you can add other team members. You can change the rights they have on the items in the space - e.g. you can grant them Edit rights or View rights for this space.

Click the 16 icon to inspect the individual team members and their rights separately.

img manage space access
Figure 3. Manage the access to a space

Manage team members

You can manage your team by going to the team’s profile. To do so, click your profile icon in the top right corner of KNIME Hub.

img teamowner profile icon
Figure 4. Select a team to go to the team’s profile page

In the dropdown menu that opens, you will see your teams. Select the team you want to manage to go to the corresponding team’s profile page.

img team profile page
Figure 5. The team’s profile page

Here, you can click the Manage team button to open the Manage members side panel, as shown in the image below.

img manage team members
Figure 6. Manage team members

You will see here a list of the team members and their assigned roles.

From here, a team admin can change the roles of the team members. To do so, click the drop down arrow close to the name and select the roles you want to assign to each user.

img change team members role
Figure 7. Manage team members roles

Then click the Save changes button to apply your changes.

Add members to a team

To add a new member, enter the username of the users that you want to add to the team in the search bar in the Manage members panel, then click the Save changes button to apply your changes.

When the maximum number of users allowed by purchase is reached, you will be notified with a message in the Manage members panel. Click the Manage subscription button to purchase more users. Alternatively, you can remove previously added users in order to add new ones.

Delete members from a team

To delete a member, go to the Manage members panel and click the 16 icon for the user you want to delete. Then click the Save changes button to apply your changes.

Change team name

On the team’s profile page, you can also change the name of the team. To do so, double-click the name of the team under the team logo on the left side of the page.

img change team name
Figure 8. Change name of the team

Insert the new name, and click the 16 button to confirm.

Change team profile icon

On the team’s profile page, hover over the team profile icon and click on Upload new to select an image from your local computer.

Manage team subscription

You can manage your subscription on the Subscription page. On the team’s profile page, select Subscription on the side menu to access it.

img change team settings
Figure 9. Manage team subscription

Here you can:

*Show the cost breakdown: Click Show cost breakdown to see the amount to be paid for your Team plan and the consumed credits for execution.

  • Manage subscription: Click Manage subscription and change the number of Users or the amount of Disk storage available for your team.

  • End subscription: In the Manage subscription side panel, click End your subscription now. This action will take effect after the end of your current billing cycle. Your data will be kept for 30 days and you will be able to re-activate your account in that time period. After 30 days we will delete all your data so it is not recoverable anymore. Once you connfirm with Cancel subscription, your team and its spaces will be deleted.

    To re-activate your Team plan please get in contact with us.

  • Past invoices: Here you can see an overview of all the invoices and download them by clicking Download invoice.

  • Inspect your billing information and edit your payment details.

Execution of workflows

Execution allows you to run and schedule your team’s workflows directly on KNIME Hub.

To do this the first step is to configure the resources you need.

As the team admin, go to your team’s page and select Execution resources from the menu on the left. Here you can create an execution context, which contains all the necessary resources to run your workflows.

This is how your page will look like the first time you create an execution context.

img start exec context
Figure 10. Start by creating an execution context

When you click the button to create a new execution context a panel on the right will open.

img create execution context
Figure 11. Create an execution context for your team

Here you can configure the capacity of your execution context. Select the size of the execution context to decide how performant your workflows execution will be. You and your team members can execute as many workflows as you wish, even in parallel.

We offer three different execution context capacities:

  • Small: Small has 2vCores and 8GB of RAM. This option is suitable for lightweight data processing with low complexity workflows. The execution power would be slightly slower than executing the workflow on your local KNIME Analytics Platform installation, so we recommend this option if you want to perform lightweight automation, e.g. schedule the execution of your workflows.

  • Medium: Medium has 4vCores and 16GB of RAM. This option is suitable for most data processing use cases with medium complex workflows. The performance is similar to a local KNIME Analytics Platform installation.

  • Large: Large had 8vCores and 32GB of RAM. If you have more demanding use cases, with heavier data, complex workflows and you need more parallel processing power with faster computation times you can select this option.

To have more information on the pricing go the Pricing section.

You will be charged only when the execution context is running. The execution context will start and stop automatically on demand. This means that it will start automatically when the execution of a workflow starts. The execution context will then stop automatically once all the workflows executions are finished.

Each execution context is based on a specific version of the KNIME Analytics Platform provided by us. Find an updated list of the extensions available on the executor here.

img ec version
Figure 12. The executor version provided by us

Manage execution contexts

You can manage the execution context of your team by navigating to your team overview page and selecting Execution resources from the menu on the left.

Here, among other things, you can edit the execution context settings, disable or delete your execution context, or show the details of the CPU usage, the jobs that are running on the execution context and so on.

To do so select the appropriate option in the menu that opens when clicking 16 in the tile of the execution context.

Monitor execution contexts consumption

To monitor your execution consumption and projected costs, you can go to your subscription page.

Go to your team page and select Subscription from the menu on the left.

img subscription with execution
Figure 13. Subscription overview

Execute a workflow

Once the execution context is set up you can run and schedule the workflows that are uploaded to your team’s spaces on KNIME Community Hub. Navigate to a workflow in one of your team’s spaces, you will see a Run button and a Deploy button.

Click Run to simply execute the current workflow directly in the browser. The latest version of your workflow will be executed, or you can select which version of the workflow you want to run. In the side panel that opens you can also enable the workflow actions, for example allowing the system to send you an email when the chose condition is met.

Be aware that this action will start up your executor, which, once the execution is started will start the billing.
img ad hoc execution panel
Figure 14. The ad hoc execution side panel

A new tab will open where the result of the workflow execution will be shown. If your workflow is a data app you will be able to interact with its user interface. To know more about how to build a data app refer to the KNIME Data Apps Beginners Guide.

To see all the ad hoc jobs created for a workflow navigate to the workflow page and select Ad hoc jobs from the right side menu.

Schedule a workflow execution

A schedule allows you to set up a workflow to run automatically at selected times.

To create a schedule deployment of your workflow navigate to a workflow in one of your team’s spaces. In order to be able to create a schedule deployment you need to have created at least one version of your workflow. To know more about versioning read the Versioning section of this guide. Once you have your stable version of the workflow, create a version (click VersionsCreate version). Now you can click the button Deploy and select Create schedule.

This will open a side panel where you can set up the schedule deployment.

img create schedule
Figure 15. The create schedule side panel

Here you can select a name for your schedule deployment and which version of the workflow you want to schedule for execution.

img create schedule options
Figure 16. The create schedule side panel: schedule options section.
To see all the deployments created for a workflow navigate to the workflow page and select Deployments from the right side menu. To see all the deployments created by your team navigate to the team page and select Deployments from the left side menu.

In the section Schedule options you can define when the workflow should be executed:

  • Set up an initial execution, date and time

  • Select Repeat execution to select if you want the workflow to run every specific amount of time (e.g. every 2 hours), or select start times when your workflow will be executed every day at the same time.

  • Set up also a schedule ends date, either to never end or to end in a specific date and time.

Set additional schedule details

You can also set more advanced schedule details.

First, check the option Repeat execution, under Set schedule details, then you can click Set to set up recurring executions, retries, and advanced schedule details.

Click Set and the following panel will open:

img schedule details
Figure 17. The create schedule side panel: set schedule details section.
If you did not check the option Repeat execution, you will only find the set up options for retries and the advanced schedule details.

The workflow will run when all the selected conditions are met. In the above example the workflow will run from Monday to Friday, every day of the month, except for the month of December.

Finally, in the section execution retries and advanced schedule details you can set up the number of execution retries, and check the following options for the advanced schedule details:

  • Reset before execution: the workflow will be reset before each scheduled execution retries occur.

  • Skip execution if previous job is still running: the scheduled execution will not take place if the previous scheduled execution is still running.

  • Disable schedule: Check this option to disable the schedule. The scheduled execution will start run accordingly to the set ups when it is re-enabled again.

Set advanced settings

Finally, you can select advanced settings for each schedule deployment. Click Set under Advanced settings section to do so.

In the panel that appears you can configure additional set ups:

  • Job lifecycle: such as deciding in which case to discard a job, the maximum time a job will stay in memory, the job life time, or the options for timeout. Particularly important in this section is the parameter Max job execution time: the default setting can be changed according to your needs to keep control over the execution time.

  • Additional settings: such as report timeouts, CPU and RAM requirements, check the option to update the linked components when executing the scheduled job and so on

The values in the fields related to time need the format:

  • PT, which stands for Period of Time

  • the amount of time, made of a number and a letter D, H, M and S for days, hours, minutes and seconds.

For example, PT1H means a period of time of 1 hour, PT168H means a period of time of 168 hours (which is equivalent of 7 days).

Jobs

Every time a workflow is executed a job is created on KNIME Hub.

To see the list of all the jobs that are saved in memory for the ad hoc execution of a specific workflow go to the workflow page and on the right side menu click Ad hoc jobs.

To see the list of all the jobs that are saved in memory for each of the deployments created for a specific workflow, go to the workflow page and on the right side menu click Deployments. You can expand each deployment with the 16 icon on the left of the deployment.

Also you can go to your team page and find a list of all deployments created within your team. Also here you can click the 16 icon corresponding to a specific deployment to see all its jobs.

On each job you can click the 16 icon on the right of the corresponding job line in the list and perform the following operations:

  • Open: For example you can open the job from a data app and look at the results in a new tab

  • Save as workflow: You can save the job as a workflow in a space

  • Inspect: A job viewer opens in a new tab. Here, you can investigate the status of jobs.

  • Delete: You can delete the job

  • Download logs: You can download the log files of a job - this feature allows for debugging in case the execution of a workflow did not work as expected.

Inspect an executed workflow

You can inspect the status of an executed workflow, for example if the workflow execution did not succeed, or the execution is taking a long time. Click the 16 icon for the desired job and select Inspect.

img inspect job chub
Figure 18. Select Inspect from the menu to check the status of the execution

Here you can hover over info and error icons at the node’s traffic light, to visualize the message, check where the execution stopped or the status of specific nodes. You can also inspect the data at any node’s output port.

img inspect node message
Figure 19. Inspect node messages and data at the output port of a node

You can also navigate inside components and metanodes and inspect the nodes.

img inspect component
Figure 20. Navigate inside components and metanodes to inspect the incapsulated nodes

Extensions available for execution on KNIME Community Hub

All the extensions developed by KNIME are available to run workflows on KNIME Community Hub.

This includes the following extensions:

  • KNIME Expressions

  • KNIME Executor connector

  • KNIME Remote Workflow Editor for Executor

  • KNIME Remote Workflow Editor

  • KNIME Hub Additional Connectivity (Labs)

  • KNIME Hub Integration

  • KNIME Active Learning

  • KNIME AI Assistant (Labs)

  • KNIME Autoregressive integrated moving average (ARIMA)

  • KNIME Basic File System Connectors

  • KNIME Views

  • KNIME Nodes for Wide Data (many columns)

  • KNIME Big Data Connectors

  • KNIME Databricks Integration

  • KNIME Extension for Big Data File Formats

  • KNIME Extension for Apache Spark

  • KNIME Extension for Local Big Data Environments

  • KNIME Chromium Embeded Framework (CEF) Browser

  • KNIME Integrated Deployment

  • KNIME Base Chemistry Types & Nodes

  • KNIME Amazon Athena Connector

  • KNIME Amazon DynamoDB Nodes

  • KNIME Amazon Cloud Connectors

  • KNIME Amazon Machine Learning Integration

  • KNIME Amazon Redshift Connector And Tools

  • KNIME Conda Integration

  • KNIME Columnar Table Backend

  • KNIME Streaming Execution (Beta)

  • KNIME BigQuery

  • KNIME JDBC Driver For Oracle Database

  • KNIME Microsoft JDBC Driver For SQL Server

  • KNIME Vertica Driver

  • KNIME Database

  • KNIME Data Generation

  • KNIME Connectors for Common Databases

  • KNIME Distance Matrix

  • KNIME Deep Learning - Keras Integration

  • KNIME Deep Learning - ONNX Integration

  • KNIME Deep Learning - TensorFlow Integration

  • KNIME Deep Learning - TensorFlow 2 Integration

  • KNIME Email Processing

  • KNIME Ensemble Learning Wrappers

  • KNIME Expressions

  • KNIME Azure Cloud Connectors

  • KNIME Box File Handling Extension

  • KNIME Chemistry Add-Ons

  • KNIME External Tool Support

  • KNIME H2O Snowflake Integration

  • KNIME H2O Machine Learning Integration

  • KNIME H2O Machine Learning Integration - MOJO Extension

  • KNIME Extension for MOJO nodes on Spark

  • KNIME H2O Sparkling Water Integration

  • KNIME Itemset Mining

  • KNIME Math Expression (JEP)

  • KNIME Indexing and Searching

  • KNIME MDF Integration

  • KNIME Office 365 Connectors

  • KNIME SharePoint List

  • KNIME Open Street Map Integration

  • KNIME SAS7BDAT Reader

  • KNIME Excel Support

  • KNIME Power BI Integration

  • KNIME Tableau Integration

  • KNIME Textprocessing

  • KNIME Twitter Connectors

  • KNIME External Tool Support (Labs)

  • KNIME Google Connectors

  • KNIME Google Cloud Storage Connection

  • KNIME Javasnippet

  • KNIME Plotly

  • KNIME Quick Forms

  • KNIME JavaScript Views

  • KNIME JavaScript Views (Labs)

  • KNIME JSON-Processing

  • KNIME Extension for Apache Kafka (Preview)

  • KNIME Machine Learning Interpretability Extension

  • KNIME MongoDB Integration

  • KNIME Neighborgram & ParUni

  • KNIME Network Mining distance matrix support

  • KNIME Network Mining

  • KNIME Optimization extension

  • KNIME Python Integration

  • KNIME Interactive R Statistics Integration

  • KNIME Report Designer (BIRT)

  • KNIME Reporting

  • KNIME REST Client Extension

  • KNIME Salesforce Integration

  • KNIME SAP Integration based on Theobald Xtract Universal

  • KNIME Git Nodes

  • KNIME Semantic Web

  • KNIME Snowflake Integration

  • KNIME Statistics Nodes

  • KNIME Statistics Nodes (Labs)

  • KNIME Timeseries nodes

  • KNIME Timeseries (Labs)

  • KNIME Modern UI

  • KNIME Parallel Chunk Loop Nodes

  • KNIME Weak Supervision

  • KNIME Webanalytics

  • KNIME XGBoost Integration

  • KNIME XML-Processing

  • SmartSheet extension

  • KNIME AI Extension (Labs)

  • KNIME Web Interaction (Labs)

  • KNIME Nodes for Scikit-Learn (sklearn) Algorithms

  • Geospatial Analytics Extension for KNIME

  • KNIME Giskard Extension

  • KNIME Presidio Extension

  • RDKit Nodes Feature

  • Vernalis KNIME Nodes

  • Slack integration

  • Continental Nodes for KNIME

  • Lhasa public plugin

To know more about the different extensions see the Extensions and Integrations Guide.

Access data from different sources

This section provides information on how to build your workflows to connect to different data sources and which kind of authentication method is recommended for a particular data source, to then run the workflows ad hoc or schedule them on a Team plan on KNIME Community Hub.

Access data from Google Cloud

Refer to the KNIME Google Cloud Integration User Guide to know how to create an API Key and to learn more about the general usage of the Google Authenticator node.

To access data from Google Cloud the first step when building your workflow is to use a Google Authenticator node. One of the possibilities is then to use the API Key authentication method. When using the API Key authentication method in the Google Authenticator node, the API key file will need to be stored somewhere where it is accessible to the node in order for the execution to be successful.
For example, if your folder structure on the KNIME Community Hub (KCH) looks something like the one below, meaning that your API Key file is stored in the following path:

<space> -> <folder> -> <api-key.fileformat>
image11
Figure 21. Example folder structure for API Key file storage

then, you can access the API key using either one of the following paths in the Google Authenticator node configuration dialog.

  • Mountpoint relative:

    knime://knime.mountpoint//<folder>/<api-key.fileformat>
    image1
    Figure 22. Configuration dialog of Google Authenticator for mountpoint relative path
  • Workflow relative:

    knime://knime.workflow/../<api-key.fileformat>
    image2
    Figure 23. Configuration dialog of Google Authenticator for workflow relative path
  • Workflow Data Area:

    knime://knime.workflow/data/<api-key.fileformat>

    To use this path, you need to make sure you have stored the JSON or p12 file in the workflow’s Data folder.
    If the folder does not exist, create one, store the file inside, and then upload the workflow to the KNIME Community Hub.

    image5
    Figure 24. Configuration dialog of Google Authenticator for workflow data area path

Access data from Google BigQuery

In order to have access to data on Google BigQuery, the first step is to authenticate with Google, using the Google Authenticator node as described above. Then you can connect to Google BigQuery using the Google BigQuery Connector node.

However, due to license restrictions, you first need to install the latest version of the Google Big Query JDBC driver, which you can download here.

Once you have downloaded the driver, extract it into a directory and register it in your KNIME Analytics Platform as described here.

When registering the driver make sure to use bigquery_154 as ID and bigquery as the database type of the driver since it needs to match with the ID of the driver in the Community Hub Executor. So your driver registration dialog should look like this, where C:\KNIME\JDBC\BigQuery\SimbaJDBCDriverforGoogleBigQuery42_1.5.4.1008 is the directory that contains all the JAR files.

image8
Figure 25. Configuration dialog to register a new database driver

Access data from Azure

Refer to the KNIME Azure Integration User Guide for setting up the Azure integration with KNIME.

To access data from Azure, the first step when building your workflow is to use a Microsoft Authenticator node.

When configuring the node, select Azure Storage Shared Key as authentication type.

Enter the Storage Account Name and Shared/Access key and connect to Azure.

image9
Figure 26. Configuration dialog of Microsoft Authenticator node

The credentials could be stored inside the node, which is not recommended, or passed to the node using the Credentials Configuration node. This will allow the team admin to pass the credentials while deploying the workflow, as shown in the below snapshot.

image3
Figure 27. Create a schedule deployment using credentials

Once authenticated, use Azure Data Lake Storage Gen2 Connector or Azure Blob Storage Connector nodes and select the working directory from where you want to access the data.

image6
Figure 28. Configuration dialog of Azure Data Lake Storage Gen2 Connector node
image7
Figure 29. Configuration dialog of Azure Blob Storage Connector node

Enter the file name in the respective file reader node and read the file.

image10
Figure 30. Configuration dialog of the File Reader node

Access data from Snowflake

Refer to the KNIME Snowflake Extension Guide for setting up Snowflake with KNIME.

To access data from Snowflake, the first step when building your workflow is to use a Snowflake Connector node.

The credentials could be stored inside the node, which is not recommended, or passed to the node using the Credentials Configuration node. This will allow the team admin to pass the credentials while deploying the workflow.

image3
Figure 31. Create a schedule deployment using credentials