tSnowflakeOutput properties for Apache Spark Batch (technical preview)
These properties are used to configure tSnowflakeOutput running in the Spark Batch Job framework.
The Spark Batch
tSnowflakeOutput component belongs to the
Databases family.
The component in this framework is available in all subscription-based Talend products with Big Data
and Talend Data Fabric.
Basic settings
Use an existing configuration |
Select this check box and in the Component List click the relevant connection component to |
Account |
In the Account field, enter, in double quotation marks, the account name |
Snowflake Region |
Select an AWS region or an Azure region from |
User Id and Password |
Enter, in double quotation marks, your authentication
|
Warehouse |
Enter, in double quotation marks, the name of the |
Schema |
Enter, within double quotation marks, the name of the |
Database |
Enter, in double quotation marks, the name of the |
Table |
Click the […] button and in the displayed wizard, select the Snowflake |
Schema and Edit Schema |
A schema is a row description. It defines the number of fields Built-In: You create and store the schema locally for this component Repository: You have already created the schema and stored it in the If the Snowflake data type to Click Edit
Note that if the input value of any non-nullable primitive |
Output Action |
Only the Insert action is supported by Snowflake on Spark. |
Usage
Usage rule |
This component is used as an end component and requires an input link. Use a tSnowFlakeConfiguration: update component in the same Job to connect |
Spark Connection |
In the Spark
Configuration tab in the Run view, define the connection to a given Spark cluster for the whole Job. In addition, since the Job expects its dependent jar files for execution, you must specify the directory in the file system to which these jar files are transferred so that Spark can access these files:
This connection is effective on a per-Job basis. |