August 15, 2023

tHConvertFile – Docs for ESB 6.x

tHConvertFile

Uses
Talend
Data Mapper
structures to perform a conversion from one representation to another, as a Spark
Batch execution.

tHConvertFile converts data from one
representation to another in a Spark environment.

tHConvertFile properties for Apache Spark Batch

These properties are used to configure tHConvertFile running in the Spark Batch Job framework.

The Spark Batch
tHConvertFile component belongs to the Processing family.

This component is available in the Palette of the Studio only if you have subscribed to any Talend Platform product with Big Data or Talend Data Fabric.

Basic settings

Storage

To connect to an HDFS installation, select the Define a storage configuration component check box and then
select the name of the component to use from those available in the
drop-down list.

This option requires you to have previously configured the connection
to the HDFS installation to be used, as described in the documentation
for the tHDFSConfiguration
component.

If you leave the Define a storage
configuration component
check box unselected, you can only
convert files locally.

Configure Component

To configure the component, click the […] button and, in the [Component
Configuration]
window, perform the following actions.

  1. Click the Select button
    next to the Record structure
    field and then, in the [Select a
    Structure]
    dialog box that opens, select the
    structure you want to use when converting your file and then
    click OK.

    This structure must have been previously created in

    Talend Data Mapper
    .

  2. Select the Input
    Representation
    to use from the drop-down
    list.

    Supported input formats are Avro, COBOL, EDI,
    Flat, IDocs, JSON and XML.

  3. Select the Output
    Representation
    to use from the drop-down list. The
    choices available for the Output representation depend on what
    you choose for the Input representation.

    Supported output formats are Avro, Flat, JSON and XML.

  4. Click Next.

  5. Tell the component where each new record begins.
    In order for you to be able to do so, you need to fully
    understand the structure of your data.

    Exactly how you do this varies depending on the
    input representation being used, and you will be presented with
    one of the following options.

    1. Select an appropriate record delimiter
      for your data. Note that you must specify this value
      without quotes.

      • Separator lets you specify a
        separator indicator, such as
        , to identify a new line.

        Supported indicators are
        for a Unix-type new line,

        for Windows and
        for Mac, and for tab characters.

      • Start/End with lets you specify the
        initial characters that indicate a new record,
        such as <root, or the
        characters that indicate where a record ends. This
        can also be a regular expression.

        Start
        with
        also supports new lines,
        for a Unix-type new line,

        for Windows and
        for Mac, and for tab characters.

      • Sample File: To test the
        signature with a sample file, click the
        […] button, browse to the
        file you want to use as a sample, click
        Open, and then click
        Run to test your
        sample.

        Testing the signature lets you check that
        the total number of records and their minimum and
        maximum length corresponds to what you expect based
        on your knowledge of the data. This step assumes you
        have a local subset of your data to use as a
        sample.

      • Click Finish.

    2. If your input representation is COBOL or
      Flat with positional and/or binary encoding properties,
      define the signature for the input record structure:

      • Input Record root corresponds
        to the root element in your input record.
      • Minimum Record Size
        corresponds to the size in bytes of the smallest
        record. If you set this value too low, you may
        encounter performance issues, since the component
        will perform more checks than necessary when
        looking for a new record.

      • Maximum Record Size
        corresponds to the size in bytes of the largest
        record, and is used to determine how much memory
        is allocated to read the input.

      • Sample from Workspace or
        Sample from File System: To
        test the signature with a sample file, click the
        […] button,
        and then browse to the file you want to use as a
        sample.

        Testing the signature lets you check that
        the total number of records and their minimum and
        maximum length corresponds to what you expect based
        on your knowledge of the data. This step assumes you
        have a local subset of your data to use as a
        sample.

      • Footer
        Size
        corresponds to the size in bytes
        of the footer, if any. At runtime, the footer will
        be ignored rather than being mistakenly included
        in the last record. Leave this field empty if
        there is no footer.

      • Click the Next button to open
        the [Signature
        Parameters]
        window, select the fields
        that define the signature of your record input
        structure (that is, to identify where a new record
        begins), update the Operation and Value columns as
        appropriate, and then click Next.

      • In the [Record
        Signature Test]
        window that opens,
        check that your Records are correctly delineated
        by scrolling through them with the
        Back and
        Next buttons and performing
        a visual check, and then click
        Finish.

Input

Click the […] button to define the path to where the input
file is stored.

You can also enter the path manually, between quotes.

Output

Click the […] button to define the path to where the output
file will be stored.

You can also enter the path manually, between quotes.

Action

From the drop-down list, select:

  • Create if you want the
    conversion process to create a new file.

  • Overwrite if you want the
    conversion process to overwrite an existing file.

Open Structure Editor

Click the […] button to open the
structure for editing in the Structure
Editor
of
Talend Data Mapper
.

For more information, see
Talend Data Mapper User
Guide
.

Advanced settings

Die on error

This check box is selected by default. Clear the check box to skip
the row in error and complete the process for error-free rows. If
needed, you can retrieve the rows in error via a Row > Rejects link.

Usage

Usage rule

This component is used with a tHDFSConfiguration component which defines the
connection to the HDFS storage, or as a standalone component for
converting local files only.

Related scenarios


Document get from Talend https://help.talend.com
Thank you for watching.
Subscribe
Notify of
guest
0 Comments
Inline Feedbacks
View all comments
0
Would love your thoughts, please comment.x
()
x