Job Example 11: Loading Hadoop Tables Using the TDCH-TPT Interface - Parallel Transporter

Teradata Parallel Transporter User Guide

Product
Parallel Transporter
Release Number
16.10
Published
May 2017
Language
English (United States)
Last Update
2018-05-09
dita:mapPath
vyv1488824663502.ditamap
dita:ditavalPath
Audience_PDF_product_tpt_userguide_include.ditaval
dita:id
B035-2445
lifecycle
previous
Product Category
Teradata Tools and Utilities

Job Objective

Use a TPT script with multiple job steps to accomplish the following tasks:

  • Create a new table in the Teradata Database.
  • Read data from a Hadoop table and write data to the newly created Teradata Database table.

Data Flow Diagram

The following figure shows a flow diagram of the elements in Job Example 11.

Job Example PTS00030 – Loading Hadoop Tables Using the TDCH-TPT Interface

Sample Script

For the sample script that corresponds to this job, see the following script in the sample/userguide directory:

PTS00030: Loading Hadoop Tables Using the TDCH-TPT Interface.

Rationale

In the first job step, this script uses:

  • DDL operator because it can DROP/CREATE tables prior to loading.

In the second job step, this script uses:

  • DataConnector operator as the producer operator because it can read data from Hadoop tables via the TDCH-TPT interface.
  • Load operator template as the consumer because it is the consumer operator that offers the best performance for high speed writing of a large number of rows into an empty Teradata Database table.