August 16, 2023

Scenario: Deduplicating entries using Map/Reduce components – Docs for ESB 6.x

Scenario: Deduplicating entries using Map/Reduce components

This scenario applies only to a subscription-based Talend Platform solution with Big data or Talend Data Fabric.

This scenario illustrates how to create a
Talend
Map/Reduce Job to
deduplicate entries, that is to say, to use Map/Reduce components to generate Map/Reduce
code and run the Job right in Hadoop.

use_case-mr_tuniqrow1.png

Note that the
Talend
Map/Reduce components are available to
subscription-based Big Data users only and this scenario can be replicated only with
Map/Reduce components.

The sample data to be used in this scenario reads as follows:

Since
Talend Studio
allows you to convert a Job between its
Map/Reduce and Standard (Non Map/Reduce) versions, you can convert the scenario
explained earlier to create this Map/Reduce Job. This way, many components used can keep
their original settings so as to reduce your workload in designing this Job.

Before starting to replicate this scenario, ensure that you have appropriate rights
and permissions to access the Hadoop distribution to be used. Then proceed as
follows:


Document get from Talend https://help.talend.com
Thank you for watching.
Subscribe
Notify of
guest
0 Comments
Inline Feedbacks
View all comments
0
Would love your thoughts, please comment.x
()
x