August 16, 2023

tSetKeystore properties for Apache Spark Batch – Docs for ESB 6.x

tSetKeystore properties for Apache Spark Batch

These properties are used to configure tSetKeystore running in the Spark Batch Job framework.

The Spark Batch
tSetKeystore component belongs to the Authentication family.

The component in this framework is available only if you have subscribed to one
of the
Talend
solutions with Big Data.

Basic settings

TrustStore type

Select the type of the TrustStore to be used. It may be PKCS 12 or JKS.

TrustStore file

Type in the path, or browse to the certificate TrustStore file
(including filename) that contains the list of certificates that the
client trusts.

TrustStore password

Type in the password used to check the integrity of the TrustStore
data.

Need Client authentication

Select this check box to validate the keystore data. Once doing so,
you need complete three fields:

KeyStore type: select the type of
the keystore to be used. It may be PKCS
12
or JKS.

KeyStore file: type in the path,
or browse to the file (including filename) containing the keystore
data.

KeyStore password: type in the
password for this keystore.

Check server identity

Select this check box to make the Job verify the match between the hostname
of the URL and the hostname of the server. If they mismatch, the verification
mechanism asks whether this connection should be allowed.

Usage

Usage rule

This component is used with no need to be connected to other components.

This component, along with the Spark Batch component Palette it belongs to, appears only
when you are creating a Spark Batch Job.

Note that in this documentation, unless otherwise
explicitly stated, a scenario presents only Standard Jobs,
that is to say traditional
Talend
data integration Jobs.

Spark Connection

You need to use the Spark Configuration tab in
the Run view to define the connection to a given
Spark cluster for the whole Job. In addition, since the Job expects its dependent jar
files for execution, you must specify the directory in the file system to which these
jar files are transferred so that Spark can access these files:

  • Yarn mode: when using Google
    Dataproc, specify a bucket in the Google Storage staging
    bucket
    field in the Spark
    configuration
    tab; when using other distributions, use a
    tHDFSConfiguration
    component to specify the directory.

  • Standalone mode: you need to choose
    the configuration component depending on the file system you are using, such
    as tHDFSConfiguration
    or tS3Configuration.

This connection is effective on a per-Job basis.


Document get from Talend https://help.talend.com
Thank you for watching.
Subscribe
Notify of
guest
0 Comments
Inline Feedbacks
View all comments
0
Would love your thoughts, please comment.x
()
x