July 30, 2023

tJDBCOutput Storm properties (deprecated) – Docs for ESB 7.x

tJDBCOutput Storm properties (deprecated)

These properties are used to configure tJDBCOutput running in the Storm Job

The Storm
tJDBCOutput component belongs to the Storm and the Databases families.

This component is available in Talend Real Time Big Data Platform and Talend Data Fabric.

The Storm framework is deprecated from Talend 7.1 onwards. Use Talend Jobs for Apache Spark Streaming to accomplish your Streaming related tasks.

Basic settings

Property type

Either Built-In or Repository.


Built-In: No property data stored centrally.


Repository: Select the repository file where the
properties are stored.

tJDBCOutput Storm properties (deprecated)_1.png

Click this icon to open a database connection wizard and store the
database connection parameters you set in the component Basic

For more information about setting up and storing database
connection parameters, see Talend Studio User Guide.


The JDBC URL of the database to be used. For
example, the JDBC URL for the Amazon Redshift database is jdbc:redshift://endpoint:port/database.

Driver JAR

Complete this table to load the driver JARs needed. To do
this, click the [+] button under the table to add
as many rows as needed, each row for a driver JAR, then select the cell and click the
[…] button at the right side of the cell to
open the Module dialog box from which you can select the driver JAR
to be used. For example, the driver jar RedshiftJDBC41- for the Redshift database.

For more information, see Importing a database driver.

Class Name

Enter the class name for the specified driver between double
quotation marks. For example, for the RedshiftJDBC41- driver, the name to be entered is

Username and

The JDBC URL of the database to be used. For
example, the JDBC URL for the Amazon Redshift database is jdbc:redshift://endpoint:port/database.

To enter the password, click the […] button next to the
password field, and then in the pop-up dialog box enter the password between double quotes
and click OK to save the settings.

Table name

Name of the table to be written. Note that only one table can
be written at a time.

Schema and
Edit schema

A schema is a row description. It defines the number of fields
(columns) to be processed and passed on to the next component. When you create a Spark
Job, avoid the reserved word line when naming the


Built-In: You create and store the schema locally for this component


Repository: You have already created the schema and stored it in the
Repository. You can reuse it in various projects and Job designs.


Usage rule

In a
Storm Job, it is used as an end component. The other
components used along with it must be Storm components, too. They generate native Storm code
that can be executed directly in a Storm system.

This component, along with the Storm family it belongs to, appears only when you are
creating a Storm Job.

Note that in this documentation, unless otherwise explicitly stated, a scenario presents
only Standard Jobs, that is to say traditional
integration Jobs.

Storm Connection

You need to use the Storm Configuration tab in the
Run view to define the connection to a given Storm
system for the whole Job.

This connection is effective on a per-Job basis.

Document get from Talend https://help.talend.com
Thank you for watching.
Notify of
Inline Feedbacks
View all comments
Would love your thoughts, please comment.x