tFileOutputXML
schema.
Depending on the Talend
product you are using, this component can be used in one, some or all of the following
Job frameworks:
-
Standard: see tFileOutputXML Standard properties.
The component in this framework is available in all Talend
products. -
MapReduce: see tFileOutputXML MapReduce properties (deprecated).
The component in this framework is available in all subscription-based Talend products with Big Data
and Talend Data Fabric. -
Spark Batch: see tFileOutputXML properties for Apache Spark Batch.
The component in this framework is available in all subscription-based Talend products with Big Data
and Talend Data Fabric. -
Spark Streaming: see tFileOutputXML properties for Apache Spark Streaming.
This component is available in Talend Real Time Big Data Platform and Talend Data Fabric.
tFileOutputXML Standard properties
These properties are used to configure tFileOutputXML running in the Standard Job framework.
The Standard
tFileOutputXML component belongs to the File and the XML families.
The component in this framework is available in all Talend
products.
Basic settings
File Name |
Name or path to the output file and/or the variable to be used. Related topic: see Defining variables from the Component view section Warning: Use absolute path (instead of relative path) for
this field to avoid possible errors. |
Incoming record is a document |
Select this check box if the data from the preceding component is in When this check box is selected, a Column When this check box is selected, in the Advanced |
Row tag |
Specify the tag that will wrap data and structure per row. |
Schema and Edit |
A schema is a row description. It defines the number of fields Click Edit
|
 |
Built-In: You create and store the schema locally for this component |
 |
Repository: You have already created the schema and stored it in the |
Sync columns | Click to synchronize the output file schema with the input file schema. The Sync function only displays once the Row connection is linked with the input component. |
Advanced settings
Split output in several files |
If the output is big, you can split the output into several files,
Rows in each output file: Specify the |
Create directory if not exists | This check box is selected by default. It creates a directory to hold the output XML files if required. |
Root tags |
Specify one or more root tags to wrap the whole output file structure and data. The |
Output format |
Define the output format.
Note:
If the same column is selected in both the Output format table as an attribute
Use schema column name: By |
Use dynamic grouping |
Select this check box if you want to dynamically group the output
Column: Select a column you want
Attribute label: Enter an |
Custom the flush buffer size |
Select this check box to define the number of rows to buffer before
Row Number: Specify the number of |
Advanced separator (for numbers) |
Select this check box to change the separator used for numbers. By default, the thousands separator is a comma (,) and the decimal separator is a period (.).
Thousands separator: define
Decimal separator: define separators |
Encoding |
Select the encoding from the list or select Custom |
Don’t generate empty file | Select the check box to avoid the generation of an empty file. |
Trim data |
Select this check box to remove the spaces at the beginning and at the |
tStatCatcher Statistics | Select this check box to gather the Job processing metadata at a Job level as well as at each component level. |
Global Variables
Global Variables |
ERROR_MESSAGE: the error message generated by the
NB_LINE: the number of rows processed. This is an After A Flow variable functions during the execution of a component while an After variable To fill up a field or expression with a variable, press Ctrl + For further information about variables, see |
Usage
Usage rule |
Use this component to write an XML file with data passed on from other |
Related scenarios
For related scenarios using tFileOutputXML, see Reading a Positional file and saving filtered results to XML and Using a SOAP message from an XML file to get country name information and saving the information to an XML file.
tFileOutputXML MapReduce properties (deprecated)
These properties are used to configure tFileOutputXML running in the MapReduce Job framework.
The MapReduce
tFileOutputXML component belongs to the MapReduce family.
The component in this framework is available in all subscription-based Talend products with Big Data
and Talend Data Fabric.
The MapReduce framework is deprecated from Talend 7.3 onwards. Use Talend Jobs for Apache Spark to accomplish your integration tasks.
In a
Talend
Map/Reduce Job, tFileOutputXML, as well as the whole Map/Reduce Job using it, generates
native Map/Reduce code. This section presents the specific properties of tFileOutputXML when it is used in that situation. For further
information about a
Talend
Map/Reduce Job, see the
Talend Open Studio for Big Data Getting Started Guide
.
Basic settings
Property type |
Either Built-In or Repository. |
 |
Built-In: No property data stored centrally. |
|
Click this icon to open a database connection wizard and store the For more information about setting up and storing database |
 |
Repository: Select the repository file where the The properties are stored centrally under the Hadoop The fields that come after are pre-filled in using the fetched For further information about the Hadoop |
Row tag |
Specify the tag that will wrap data and structure per row. |
Schema and Edit |
A schema is a row description. It defines the number of fields Click Edit
|
 |
Built-In: You create and store the schema locally for this component |
 |
Repository: You have already created the schema and stored it in the |
Folder |
Browse to, or enter the path pointing to the data to be used in the file system. This path must point to a folder rather than a file, because a Note that you need |
Action |
Select an operation for writing data:
Create: Creates a file and write data
Overwrite: Overwrites the file |
Advanced settings
Root tags |
Specify one or more root tags to wrap the whole output file structure and data. The |
Output format |
Define the output format.
Note:
If the same column is selected in both the Output format table as an attribute
Use schema column name: By |
Use dynamic grouping |
Select this check box if you want to dynamically group the output
Column: Select a column you want
Attribute label: Enter an |
Encoding |
Select the encoding from the list or select Custom |
Use local timezone for date | Select this check box to use the local date of the machine in which your Job is executed. If leaving this check box clear, UTC is automatically used to format the Date-type data. |
Global Variables
Global Variables |
ERROR_MESSAGE: the error message generated by the A Flow variable functions during the execution of a component while an After variable To fill up a field or expression with a variable, press Ctrl + For further information about variables, see |
Usage
Usage rule |
In a Once a Map/Reduce Job is opened in the workspace, tFileOutputXML as well as the MapReduce Note that in this documentation, unless otherwise |
Hadoop Connection |
You need to use the Hadoop Configuration tab in the This connection is effective on a per-Job basis. |
Related scenarios
No scenario is available for the Map/Reduce version of this component yet.
tFileOutputXML properties for Apache Spark Batch
These properties are used to configure tFileOutputXML running in the Spark Batch Job framework.
The Spark Batch
tFileOutputXML component belongs to the File and the XML families.
The component in this framework is available in all subscription-based Talend products with Big Data
and Talend Data Fabric.
Basic settings
Define a storage configuration |
Select the configuration component to be used to provide the configuration If you leave this check box clear, the target file system is the local The configuration component to be used must be present in the same Job. |
Property type |
Either Built-In or Repository. |
 |
Built-In: No property data stored centrally. |
|
Click this icon to open a database connection wizard and store the For more information about setting up and storing database |
 |
Repository: Select the repository file where the The properties are stored centrally under the Hadoop The fields that come after are pre-filled in using the fetched For further information about the Hadoop |
Row tag |
Specify the tag that will wrap data and structure per row. |
Schema and Edit |
A schema is a row description. It defines the number of fields Click Edit
|
 |
Built-In: You create and store the schema locally for this component |
 |
Repository: You have already created the schema and stored it in the |
Folder |
Browse to, or enter the path pointing to the data to be used in the file system. This path must point to a folder rather than a file. The button for browsing does not work with the Spark tHDFSConfiguration |
Action |
Select an operation for writing data:
Create: Creates a file and write data
Overwrite: Overwrites the file |
Compress the data |
Select the Compress the data check box to compress the |
Advanced settings
Root tags |
Specify one or more root tags to wrap the whole output file structure and data. The |
Output format |
Define the output format.
Note:
If the same column is selected in both the Output format table as an attribute
Use schema column name: By |
Use dynamic grouping |
Select this check box if you want to dynamically group the output
Column: Select a column you want
Attribute label: Enter an |
Custom encoding |
Select the encoding from the list or select Custom |
Advanced separator (for numbers) |
Select this check box to modify the separators used for
Thousands separator: define
Decimal separator: define |
Use local timezone for date | Select this check box to use the local date of the machine in which your Job is executed. If leaving this check box clear, UTC is automatically used to format the Date-type data. |
Usage
Usage rule |
This component is used as an end component and requires an input link. This component, along with the Spark Batch component Palette it belongs to, Note that in this documentation, unless otherwise explicitly stated, a |
Spark Connection |
In the Spark
Configuration tab in the Run view, define the connection to a given Spark cluster for the whole Job. In addition, since the Job expects its dependent jar files for execution, you must specify the directory in the file system to which these jar files are transferred so that Spark can access these files:
This connection is effective on a per-Job basis. |
Related scenarios
No scenario is available for the Spark Batch version of this component
yet.
tFileOutputXML properties for Apache Spark Streaming
These properties are used to configure tFileOutputXML running in the Spark Streaming Job framework.
The Spark Streaming
tFileOutputXML component belongs to the File and the XML families.
This component is available in Talend Real Time Big Data Platform and Talend Data Fabric.
Basic settings
Define a storage configuration |
Select the configuration component to be used to provide the configuration If you leave this check box clear, the target file system is the local The configuration component to be used must be present in the same Job. |
Property type |
Either Built-In or Repository. |
 |
Built-In: No property data stored centrally. |
|
Click this icon to open a database connection wizard and store the For more information about setting up and storing database |
 |
Repository: Select the repository file where the The properties are stored centrally under the Hadoop The fields that come after are pre-filled in using the fetched For further information about the Hadoop |
Row tag |
Specify the tag that will wrap data and structure per row. |
Schema and Edit |
A schema is a row description. It defines the number of fields Click Edit
|
 |
Built-In: You create and store the schema locally for this component |
 |
Repository: You have already created the schema and stored it in the |
Folder |
Browse to, or enter the path pointing to the data to be used in the file system. This path must point to a folder rather than a file. The button for browsing does not work with the Spark tHDFSConfiguration |
Action |
Select an operation for writing data:
Create: Creates a file and write data
Overwrite: Overwrites the file |
Compress the data |
Select the Compress the data check box to compress the |
Advanced settings
Root tags |
Specify one or more root tags to wrap the whole output file structure and data. The |
Output format |
Define the output format.
Note:
If the same column is selected in both the Output format table as an attribute
Use schema column name: By |
Use dynamic grouping |
Select this check box if you want to dynamically group the output
Column: Select a column you want
Attribute label: Enter an |
Custom encoding |
Select the encoding from the list or select Custom |
Advanced separator (for numbers) |
Select this check box to modify the separators used for
Thousands separator: define
Decimal separator: define |
Write empty batches | Select this check box to allow your Spark Job to create an empty batch when the incoming batch is empty. For further information about when this is desirable |
Use local timezone for date | Select this check box to use the local date of the machine in which your Job is executed. If leaving this check box clear, UTC is automatically used to format the Date-type data. |
Usage
Usage rule |
This component is used as an end component and requires an input link. This component, along with the Spark Streaming component Palette it belongs to, appears Note that in this documentation, unless otherwise explicitly stated, a scenario presents |
Spark Connection |
In the Spark
Configuration tab in the Run view, define the connection to a given Spark cluster for the whole Job. In addition, since the Job expects its dependent jar files for execution, you must specify the directory in the file system to which these jar files are transferred so that Spark can access these files:
This connection is effective on a per-Job basis. |
Related scenarios
No scenario is available for the Spark Streaming version of this component
yet.