tJDBCOutput properties for Apache Spark Batch
These properties are used to configure tJDBCOutput running in the Spark Batch
Job framework.
The Spark Batch
tJDBCOutput component belongs to the Databases family.
This component can be used to write data to a RDS MariaDB, a
RDS PostgreSQL or a RDS SQLServer database.
The component in this framework is available in all subscription-based Talend products with Big Data
and Talend Data Fabric.
Basic settings
Property type |
Either Built-In or Repository. |
||
 |
Built-In: No property data stored centrally. |
||
 |
Repository: Select the repository file where the |
||
Click this icon to open a database connection wizard and store the For more information about setting up and storing database |
|||
Use an existing |
Select this check box and in the Component List click the relevant connection component to |
||
JDBC URL |
The JDBC URL of the database to be used. For If you are using Spark V1.3, this URL should contain the
authentication information, such as:
|
||
Driver JAR |
Complete this table to load the driver JARs needed. To do For more information, see Importing a database driver. |
||
Class Name |
Enter the class name for the specified driver between double |
||
Username and |
Enter the authentication information to the database you need To enter the password, click the […] button next to the Available only for Spark V1.4. and onwards. |
||
Table name |
Name of the table to be written. Note that only one table can |
||
Schema and |
A schema is a row description. It defines the number of fields |
||
 |
Built-In: You create and store the schema locally for this component |
||
 |
Repository: You have already created the schema and stored it in the |
||
Action on data |
Select an action to be performed on data of the table defined.
|
||
Die on error |
Select the check box to stop the execution of the Job when an error |
Advanced settings
Additional JDBC |
Specify additional connection properties for the database connection you are This field is not available if the Use an existing |
Left protected |
Enter the symbol reserved by the database you are using, the left part in Left protected char and the right part in Right For example, if you are using Oracle, double quotation marks (“) are reserved for object |
Additional |
This option allows you to call SQL functions to perform actions on columns, |
 |
Name: Type in the name of the schema column to be altered |
 |
SQL expression: Type in the SQL statement to be executed in |
 |
Position: Select Before, |
 |
Reference column: Type in a reference column that the |
Use field |
Select the check box for the corresponding column to customize a request,
|
Use Batch |
Select this check box to activate the batch mode for data processing. This check box is available only when |
Batch Size |
Specify the number of records to be processed in each batch. This field appears only when the Use batch mode |
Connection pool |
In this area, you configure, for each Spark executor, the connection pool used to control
|
Evict |
Select this check box to define criteria to destroy connections in the connection pool. The
|
Usage
Usage rule |
This component is used as an end component and requires an input link. This component should use a tJDBCConfiguration component present in the same Job to This component, along with the Spark Batch component Palette it belongs to, Note that in this documentation, unless otherwise explicitly stated, a |
Spark Connection |
In the Spark
Configuration tab in the Run view, define the connection to a given Spark cluster for the whole Job. In addition, since the Job expects its dependent jar files for execution, you must specify the directory in the file system to which these jar files are transferred so that Spark can access these files:
This connection is effective on a per-Job basis. |