Warning
This component will be available in the Palette of the studio on the condition that you have subscribed to
one of the Talend solutions with Big
Data.
The tBigQueryOutputBulk and tBigQueryBulkExec components are generally used together as parts of a two
step process. In the first step, an output file is generated. In the second step, this file
is used to feed a dataset. These two steps are fused together in the tBigQueryOutput component, detailed in a separate section. The advantage of
using two separate components is that the data can be transformed before it is loaded in the
dataset.
Component family |
Big Data / Google BigQuery |
|
Function |
This component transfers a given file from Google Cloud Storage to |
|
Purpose |
This component transfers given data to Google BigQuery. |
|
Basic settings |
Schema and Edit |
A schema is a row description. It defines the number of fields to be processed and passed on Since version 5.6, both the Built-In mode and the Repository mode are Click Edit schema to make changes to the schema. If the
|
|
|
Built-In: You create and store the schema locally for this |
|
|
Repository: You have already created the schema and |
Connection |
Client ID and Client |
Paste the client ID and the client secret, both created and viewable on the API Access tab To enter the client secret, click the […] button next |
Project ID |
Paste the ID of the project hosting the BigQuery service you need to use. The default ID of this project can be found in the URL of the Google API Console, or by |
|
Authorization code |
Paste the authorization code provided by Google for the access you are building. To obtain the authorization code, you need to execute the Job using this component and |
|
Dataset |
Enter the name of the dataset you need to transfer data to. |
|
Table |
Enter the name of the table you need to transfer data to. If this table does not exist, select the Create the table if it doesn’t exist check |
|
Action on data |
Select the action to be performed from the drop-down list when
|
|
Google storage configuration |
Bulk file already exists in Google |
Select this check box to reuse the authentication information for Google Cloud Storage |
Access key and Secret key |
Paste the authentication information obtained from Google for making requests to Google To enter the secret key, click the […] button next to These keys can be consulted on the Interoperable Access tab view under the Google Cloud |
|
File to upload |
When the data to be transferred to BigQuery is not stored on Google Cloud Storage, browse |
|
Bucket |
Enter the name of the bucket, the Google Cloud Storage container, that holds the data to |
|
File |
Enter the directory of the data stored on Google Cloud Storage and to be transferred to If the data is not on Google Cloud Storage, this directory is used as the intermediate |
|
Header |
Set values to ignore the header of the transferred data. For |
|
Die on error |
This check box is cleared by default, meaning to skip the row on |
|
Advanced settings |
token properties File Name |
Enter the path to, or browse to the refresh token file you need to use. At the first Job execution using the Authorization code With only the token file name entered, Talend Studio considers the directory of that token file For further information about the refresh token, see the manual of Google BigQuery. |
Set the field delimiter |
Enter character, string or regular expression to separate fields for the transferred |
|
Encoding |
Select the encoding from the list or select Custom and |
|
tStatCatcher Statistics |
Select this check box to collect the log data at the component |
|
Global Variables |
ERROR_MESSAGE: the error message generated by the A Flow variable functions during the execution of a component while an After variable To fill up a field or expression with a variable, press Ctrl + For further information about variables, see Talend Studio |
|
Usage |
This is a standalone component. |
|
Log4j |
The activity of this component can be logged using the log4j feature. For more information on this feature, see Talend Studio User For more information on the log4j logging levels, see the Apache documentation at http://logging.apache.org/log4j/1.2/apidocs/org/apache/log4j/Level.html. |
|
Limitation |
N/A |
For related topic, see Scenario: Writing data in BigQuery