July 30, 2023

tImpalaCreateTable – Docs for ESB 7.x


Creates Impala tables that fit a wide range of Impala data formats.

tImpalaCreateTable connects to the Impala database to be
used and creates an Impala table that is dedicated to data of the format you specify.

tImpalaCreateTable Standard properties

These properties are used to configure tImpalaCreateTable running in the Standard Job framework.

The Standard
tImpalaCreateTable component belongs to the Big Data family.

The component in this framework is available in all Talend products with Big Data
and in Talend Data Fabric.

Basic settings

Property type

Either Built-in or Repository.


Built-in: No property data stored


Repository: Select the repository
file in which the properties are stored. The fields that follow are
completed automatically using the data retrieved.

Use an existing connection

Select this check box and in the Component List click the relevant connection component to
reuse the connection details you already defined.

Note: When a Job contains the parent Job and the child Job, if you
need to share an existing connection between the two levels, for example, to share the
connection created by the parent Job with the child Job, you have to:

  1. In the parent level, register the database connection
    to be shared in the Basic
    view of the connection component which creates that very database

  2. In the child level, use a dedicated connection
    component to read that registered database connection.

For an example about how to share a database connection
across Job levels, see

Talend Studio
User Guide


Select the cluster you are using from the drop-down list. The options in the
list vary depending on the component you are using. Among these options, the following
ones requires specific configuration:

  • If available in this Distribution drop-down
    list, the Microsoft HD
    option allows you to use a Microsoft HD Insight
    cluster. For this purpose, you need to configure the connections to the HD
    Insightcluster and the Windows Azure Storage service of that cluster in the
    areas that are displayed. For
    detailed explanation about these parameters, search for configuring the
    connection manually on Talend Help Center (https://help.talend.com).

  • If you select Amazon EMR, find more details about Amazon EMR getting started in
    Talend Help Center (https://help.talend.com).

  • The Custom option
    allows you to connect to a cluster different from any of the distributions
    given in this list, that is to say, to connect to a cluster not officially
    supported by

  1. Select Import from existing
    to import an officially supported distribution as base
    and then add other required jar files which the base distribution does not

  2. Select Import from zip to
    import the configuration zip for the custom distribution to be used. This zip
    file should contain the libraries of the different Hadoop elements and the index
    file of these libraries.


    Exchange, members of
    community have shared some ready-for-use configuration zip files
    which you can download from this Hadoop configuration
    list and directly use them in your connection accordingly. However, because of
    the ongoing evolution of the different Hadoop-related projects, you might not be
    able to find the configuration zip corresponding to your distribution from this
    list; then it is recommended to use the Import from
    existing version
    option to take an existing distribution as base
    to add the jars required by your distribution.

    Note that custom versions are not officially supported by

    and its community provide you with the opportunity to connect to
    custom versions from the Studio but cannot guarantee that the configuration of
    whichever version you choose will be easy, due to the wide range of different
    Hadoop distributions and versions that are available. As such, you should only
    attempt to set up such a connection if you have sufficient Hadoop experience to
    handle any issues on your own.


    In this dialog box, the active check box must be kept
    selected so as to import the jar files pertinent to the connection to be
    created between the custom distribution and this component.

    For a step-by-step example about how to connect to a custom
    distribution and share this connection, see Hortonworks.

Impala version

Select the version of the Hadoop distribution you are using. The available
options vary depending on the component you are using.


Database server IP address.


Listening port number of DB server.


Fill this field with the name of the database.

Username and

DB user authentication data.

Use kerberos authentication

If you are accessing an Impala system running with Kerberos security,
select this check box and then enter the Kerberos principal of this Impala system.

  • If this cluster is a MapR cluster of the version 5.0.0 or later, you can set the
    MapR ticket authentication configuration in addition or as an alternative by following
    the explanation in Connecting to a security-enabled MapR.

    Keep in mind that this configuration generates a new MapR security ticket for the username
    defined in the Job in each execution. If you need to reuse an existing ticket issued for the
    same username, leave both the Force MapR ticket
    check box and the Use Kerberos
    check box clear, and then MapR should be able to automatically
    find that ticket on the fly.

This check box is available depending on the Hadoop distribution you are
connecting to.

Schema and Edit

A schema is a row description. It defines the number of fields
(columns) to be processed and passed on to the next component. When you create a Spark
Job, avoid the reserved word line when naming the

Click Edit
to make changes to the schema. If the current schema is of the Repository type, three options are available:

  • View schema: choose this
    option to view the schema only.

  • Change to built-in property:
    choose this option to change the schema to Built-in for local changes.

  • Update repository connection:
    choose this option to change the schema stored in the repository and decide whether
    to propagate the changes to all the Jobs upon completion. If you just want to
    propagate the changes to the current Job, you can select No upon completion and choose this schema metadata
    again in the Repository Content


Built-In: You create and store the schema locally for this component


Repository: You have already created the schema and stored it in the
Repository. You can reuse it in various projects and Job designs.

When the schema to be reused has default values that are
integers or functions, ensure that these default values are not enclosed within
quotation marks. If they are, you must remove the quotation marks manually.

You can find more details about how to
verify default values in retrieved schema in Talend Help Center (https://help.talend.com).

Table Name

Name of the table to be created.

Action on table

Select the action to be carried out for creating a table.


Select the data format to which the table to be created is

The available data formats vary depending on the version of the
Hadoop distribution you are using.

Note that when the file format to be used is PARQUET, you might be prompted to find the specific
PARQUET jar file and install it into the Studio.

  • When the connection mode to Hive is Embedded, the Job is run in your local
    machine and calls this jar installed in the Studio.

  • When the connection mode to Hive is Standalone, the Job is run in the server
    hosting Hive and this jar file is sent to the HDFS system of the cluster you are
    connecting to. Therefore, ensure that you have properly defined the NameNode URI in the
    corresponding field of the Basic

This jar file can be downloaded from Apache’s site.
You can find more details about how to install
external modules in Talend Help Center (https://help.talend.com)

Set partitions

Select this check box to add partition columns to the table to be
created. Once selecting it, you need to define the schema of the
partition columns you need to add.

Set file location

If you want to create an Impala table in a directory other than the default one, select
this check box and enter the directory in HDFS you want to use to
hold the table content.

This is typical useful when you need to create an external Impala
table by selecting the Create an external
check box in the Advanced

Use S3 endpoint

The Use S3 endpoint check box is
displayed when you have selected the Set file
check box to create an external Impala table.

Once this Use S3
check box is selected, you need to enter the following
parameters in the fields that appear:

  • S3 bucket:
    enter the name of the bucket in which you need to create the table.

  • Bucket name: enter the name of the bucket in
    which you want to store the dependencies of your Job. This bucket must
    already exist on S3.
  • Temporary resource folder: enter the
    directory in which you want to store the dependencies of your Job. For
    example, enter temp_resources to write the dependencies in
    the /temp_resources folder in the bucket.

    If this folder already exists at runtime, its contents are overwritten
    by the upcoming dependencies; otherwise, this folder is automatically

  • Access key and
    Secret key:
    enter the authentication information required to connect to the Amazon
    S3 bucket to be used.

    To enter the password, click the […] button next to the
    password field, and then in the pop-up dialog box enter the password between double quotes
    and click OK to save the settings.

Note that the format of the S3 file is S3N (S3 Native Filesystem).

Since an Impala table created in S3 is actually an external table, this Use S3 endpoint check box must be used
with the Create an external table
case being selected.

Advanced settings

Like table

Select this check box and enter the name of the Impala table you
want to copy. This allows you to copy the definition of an existing
table without copying its data.

For further information about the Like parameter, see Cloudera’s information about
Impala’s Data Definition Language.

Create an external table

Select this check box to make the table to be created an external
Impala table. This kind of Impala table leaves the raw data where it is
if the data is in HDFS.

An external table is usually the better choice for accessing
shared data existing in a file system.

For further information about an external Impala table, see Cloudera’s documentation
about Impala.

Table comment

Enter the description you want to use for the table to be

As select

Select this check box and enter the As select statement for creating an
Impala table that is based on a Select

Table properties

Add any custom Impala table properties you want to override the
default ones used by the Hadoop engine of the Studio.

tStatCatcher Statistics

Select this check box to collect log data at the component

Global Variables

Global Variables

ERROR_MESSAGE: the error message generated by the
component when an error occurs. This is an After variable and it returns a string. This
variable functions only if the Die on error check box is
cleared, if the component has this check box.

A Flow variable functions during the execution of a component while an After variable
functions after the execution of the component.

To fill up a field or expression with a variable, press Ctrl +
to access the variable list and choose the variable to use from it.

For further information about variables, see
Talend Studio

User Guide.


Usage rule

This component works standalone.

Row format

Set Delimited row format


Die on error

Dynamic settings

Click the [+] button to add a row in the table
and fill the Code field with a context
variable to choose your database connection dynamically from multiple
connections planned in your Job. This feature is useful when you need to
access database tables having the same data structure but in different
databases, especially when you are working in an environment where you
cannot change your Job settings, for example, when your Job has to be
deployed and executed independent of Talend Studio.

The Dynamic settings table is
available only when the Use an existing
check box is selected in the Basic settings view. Once a dynamic parameter is
defined, the Component List box in the
Basic settings view becomes unusable.

For examples on using dynamic parameters, see Reading data from databases through context-based dynamic connections and Reading data from different MySQL databases using dynamically loaded connection parameters. For more information on Dynamic
and context variables, see Talend Studio
User Guide.


The Hadoop distribution must be properly installed, so as to guarantee the interaction
Talend Studio
. The following list presents MapR related information for

  • Ensure that you have installed the MapR client in the machine where the Studio is,
    and added the MapR client library to the PATH variable of that machine. According
    to MapR’s documentation, the library or libraries of a MapR client corresponding to
    each OS version can be found under MAPR_INSTALL
    . For example, the library for
    Windows is lib
    in the MapR
    client jar file. For further information, see the following link from MapR: http://www.mapr.com/blog/basic-notes-on-configuring-eclipse-as-a-hadoop-development-environment-for-mapr.

    Without adding the specified library or libraries, you may encounter the following
    error: no MapRClient in java.library.path.

  • Set the -Djava.library.path argument, for example, in the Job Run VM arguments area
    of the Run/Debug view in the Preferences dialog box in the Window menu. This argument provides to the Studio the path to the
    native library of that MapR client. This allows the subscription-based users to make
    full use of the Data viewer to view locally in the
    Studio the data stored in MapR.

For further information about how to install a Hadoop distribution, see the manuals
corresponding to the Hadoop distribution you are using.

Related scenario

This component is used in the similar way as a tHiveCreateTable component is. For further information, see Creating a partitioned Hive table.

Document get from Talend https://help.talend.com
Thank you for watching.
Notify of
Inline Feedbacks
View all comments
Would love your thoughts, please comment.x