start (Teradata Database) | Commands| Teradata Data Mover - start - Teradata Data Mover

Teradata® Data Mover User Guide

Product
Teradata Data Mover
Release Number
17.05
Published
October 2020
Language
English (United States)
Last Update
2020-11-18
dita:mapPath
xxl1591023586203.ditamap
dita:ditavalPath
mpm1591127278842.ditaval
dita:id
B035-4101
lifecycle
previous
Product Category
Analytical Ecosystem

Purpose

The start command starts a job that was created with the create command. You may specify different job variable values than were originally used by entering them in the command line at runtime. You may also modify job variable values as well as the list of objects to copy by supplying an updated parameters.xml file. If the daemon does not have sufficient resources to run the job immediately, the job is queued.

Parameters

See Parameter Order.

data_streams
[Optional] Number of data streams to use between the source and target databases. Applies to jobs that use Teradata DSA and TPT API (to and from Teradata). All other protocols use a single data stream.
Example:  4
The default value is dynamically calculated by Data Mover.
db_client_encryption
[Optional] Set to true if job needs to be encrypted during data transfer.
dm.rest.endpoint
[Optional] Enter a Data Mover REST server URL to overwrite the default value specified in the commandline.properties file in order to connect to a different REST server (and therefore a different daemon) at runtime.
https://dm-server1:1443/datamover
force_utility
[Optional] Forces the Data Mover daemon to use a specific utility for all copy operations.

Valid Values

  • dsa
  • jdbc
  • tptapi
  • tptapi_load
  • tptapi_stream
  • tptapi_update
  • T2T
If this value is not specified, the Data Mover daemon determines which Teradata utility is the best to use for the job.
Copying data to an older version of the Teradata Database using Teradata DSA is not valid. You cannot use Teradata DSA if the source and target TDPIDs are the same.
Example:  dsa
job_name
Name of the job to be started.
Example:  12315DFHJKS
job_priority
[Optional] Specifies the execution priority for the job. Supported values are: HIGH, MEDIUM, and LOW, and UNSPECIFIED. If no value is specified, the default of MEDIUM is used at runtime.
Example:  MEDIUM
log_level
[Optional] Log level for log file output.

Valid Values

  • 0
  • 1
  • 2
  • 99
Example:  2
The default value is 0.
max_agents_per_task
[Optional] Maximum number of Data Mover agents to use in parallel when moving tables or databases.
Example:  4
The default value is dynamically calculated by Data Mover.
netrace
[Optional] CLI netrace parameter. Any value greater than or equal to 0 generates a CLI trace log. Valid CLI value must be provided.
netrace_buf_len
[Optional] CLI netrace_buf_len parameter. Any value greater than or equal to 0 generates a CLI trace log. Valid CLI value must be provided.
online_archive
[Optional] Allows read and write access to the source table(s) while the tables are being copied with Teradata DSA. Updates occur to the source table during the copy, but are not transferred to the target table. After a successful copy, the data contained in the target table matches the data that was in the source table at the beginning of the copy.
Valid Values
Value Description
True Enables online archive
False Disables online archive
Unspecified Default – the value is set to the value in the Data Mover daemon configuration file
Example:  true
overwrite_existing_objects
[Optional] Job overwrites objects that already exist on the target.
Valid Values
Value Description
True Enables overwriting
False Disables overwriting
Unspecified Default – the value is set to the value in the Data Mover daemon configuration file
If the parameter is not specified, the value is set to the overwrite_existing_objects parameter value in the Data Mover daemon configuration file. If the parameter is specified as true or false, that value takes precedence over the parameter value in the Data Mover daemon configuration file.
Example:  true
query_band
[Optional] A semicolon-separated set of name-value pairs that uniquely identifies Teradata sessions or transactions for the source and target. To use a query band to identify the job payroll , the user ID aa100000 , and job session number 1122, define the query band as follows:
Example:  Job=payroll;Userid=aa1000000;Jobsession=1122;
This parameter must be added as the last parameter in the XML job definition. See About Query Band for detailed syntax rules when defining a query band.
response_timeout
[Optional] Amount of time, in seconds, to wait for response from the Data Mover daemon.
Example:  60
save_changes
[Optional] Saves the changed job variable values and uses them to replace the values originally defined when the job was created.
security_password
[Optional] Password for the super user or authorized Viewpoint user.
Example:  53cUr17y
Required if security management is enabled on the Data Mover daemon. Not a valid parameter if -security_password_encrypted is also specified.
security_password_encrypted
[Optional] Encrypted password for the super user.
Example:  052c7aabd14c7770141ac3c0137ab98ae0d3f0f7cddf588981206b010c0c1b2f
Required if security management is enabled on the Data Mover daemon. Not a valid parameter if -security_password is also specified.
security_username
[Optional] User ID of the super user or authorized Viewpoint user. The user ID of the super user is dmcl_admin and cannot be changed.
Required if security management is enabled on the Data Mover daemon.
source_account_id
[Optional] Logon account ID for source database.
Spaces in the account name for the source or target account id causes the job to fail.
source_logon_mechanism
[Optional] Logon mechanism for source system. To log on to a source Teradata Database system, the user must provide at least one of the following:
  • source_user and source_password
  • source_logon_mechanism

Logon mechanisms are not supported for Teradata DSA jobs. Use logon mechanisms only for Teradata PT API and Teradata JDBC jobs. If -source_logon_mechanism is specified and -force_utility is not used, Teradata PT API is used by default. Specifying -source_logon_mechanism with Teradata DSA specified for -force_utility results in an error.

Example:  KRB5
source_logon_mechanism_data
[Optional] Additional parameters needed for the logon mechanism of the source system.
Example:  joe@domain1@@mypassword
source_password
[Optional] Source Teradata logon password.
Example:  123456789
Not a valid parameter if -source_password_encrypted is also specified. If you do not specify a password for this parameter, the command prompts you to enter it interactively. Input is masked with a set number of asterisks, regardless of the length of the password.
source_sessions
[Optional] Number of sessions per data stream on the source database.
Example:  4
The default value is dynamically calculated by Data Mover.
source_tdpid
[Optional] Source Teradata Database.
Example:  Checks
source_user
[Optional] Source Teradata logon id.
Example:  TD_API_user
If you do not specify a logon id for this parameter, the command prompts you to enter it interactively.
Spaces in the user name for the source or target id causes the job to fail.
sync
[Optional] Waits for the job to complete, then returns an exit code that indicates if the job completed successfully. An exit code of 0 indicates successful job completion. An exit code other than 0 indicates an error with the job or with the command.
target_account_id
[Optional] Logon account ID for target database.
Spaces in the account name for the source or target account id causes the job to fail.
target_logon_mechanism
[Optional] Logon mechanism for target system. To log on to a target Teradata Database system, the user must provide at least one of the following:
  • target_user and target_password
  • target_logon_mechanism

Teradata DSA does not support logon mechanisms. Use logon mechanisms only with Teradata PT API and Teradata JDBC jobs. If -target_logon_mechanism is specified and -force_utility is not used, Teradata PT API is used by default. Specifying -target_logon_mechanism with Teradata DSA specified for -force_utility results in an error.

Example:  KRB5
target_password
[Optional] Target Teradata logon password.
Example:  212133344
Not a valid parameter if -target_password_encrypted is also specified. If you do not specify a password for this parameter, the command prompts you to enter it interactively. Input is masked with a set number of asterisks, regardless of the length of the password.
target_sessions
[Optional] Number of sessions per data stream on the target database.
Example:  4
The default value is dynamically calculated by Data Mover.
target_tdpid
[Optional] Target Teradata Database.
Example:  Leo
target_user
[Optional] Target Teradata logon id.
Example:  TD_tar_User
If you do not specify a logon id for this parameter, the command prompts you to enter it interactively.
Spaces in the user name for the source or target id causes the job to fail.
tpt_debug
[Optional] TPT API trace debug log parameter. Any value greater than or equal to 0 generates a TPT API trace log. Valid TPT API value must be provided.
uowid
[Optional] Alternate ID or name for the batch of work associated with the job. If you provide a value for this parameter, Data Mover reports this value as the unit of work ID when sending events to Teradata Ecosystem Manager or to its internal TMSMEVENT table. If you do not specify this parameter, Data Mover uses a default value as the unit of work ID when sending events to Teradata Ecosystem Manager or to its internal TMSMEVENT table. The default value for the unit of work ID is composed of the job execution name and the current timestamp. For example, if you want to define the origins of a query source the job execution name is sales_table, the default value of the unit of work ID is sales_table-20110115155656EST
Example:  sales_tables_start

Usage Notes

Each time a job starts, a new job instance is created with the name job name-date time year. This name appears in the status command or the logs to identify information for this specific execution of the job. Use the status command to monitor the job status after the job has been started.

If an instance of a job has not completed all specified steps, a new instance of the job cannot be started. If the daemon does not have adequate resources to run the job immediately, the job is placed in a queue. Copying the same objects to the same target as another job that is running or queued is not allowed.

You can supply new values for job variables at runtime using either of the following two methods:
  • Specify the value for the job variable directly in the command line, just as you can using the create command. For example to set the logging level for job1 to a value of 99, type: datamove start -job_name job1 -log_level 99.
  • Specify the new value for the job variable by modifying the XML file and submitting it to the start command, just as you can using the create command: datamove start -job_name job1 -f job1.xml

If you want to modify the list of objects to copy, you must do so by modifying and submitting the XML file. Be sure that the XML file contains all of the objects you want to copy, not only those with name changes. If an object that was originally specified to be copied is no longer included in the list, it is not included in the job execution.

By default, any job variables supplied at runtime are not saved to the job definition. Therefore, if you run the job again, the job reverts to the original settings. You can save the new job variable values for future use by using the save changes parameter in either the command line or file. For example:
  • At the command line prompt, type:datamove start -job_name job1 -log_level 99 -save_changes
  • In the XML file, specify: <saveChanges>true</saveChanges>
    The save changes parameter syntax is different when used at the command line versus in an XML file, as shown in the example.
If save changes is enabled, the modified job definition is saved as the new base job definition. The modification of job variable values triggers a rebuild of the job plan, overriding the freeze job steps option, if enabled. The new job plan is then used for the current run. If save changes is not enabled, the base job definition is not changed. The job plan is rebuilt and the modified variables are used for the current run, but the base job definition remains unchanged and all subsequent executions use the base job definition.
Security settings impact certain start command options in several ways:
  • If security is enabled, a user must have write permission to be able to set the save changes parameter to true.
  • A user who is not the job owner must supply source and target Teradata system user names and passwords when specifying the object list in the XML and submitting it using the start command.
  • If security is enabled and job_security is specified in the modified XML to change job permission, the user must be dcml_admin or the job owner, and the user must provide all the permissions, not just the modified permissions. If job_owner is specified in job_security and the user wants to change the job owner, the user must be dcml_admin.

XML File Example

For the start command, type datamove start -f parameters.xml.

In the following example, the XML file shown defines the job testStart which copies tables fmt_inf and test 1 to the target using DSA.
<?xml version="1.0" encoding="UTF-8" standalone="yes" ?> 
<dmCreate xmlns="http://schemas.teradata.com/dataMover/v2009" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://schemas.teradata.com/unity/datamover.xsd"> 
    <job_name>testStart</job_name> 
    <source_tdpid>dmdev</source_tdpid> 
    <source_user>dmuser</source_user> 
    <source_password>dbc</source_password> 
    <target_tdpid>dm-daemon2</target_tdpid> 
    <target_user>dmuser</target_user> 
    <target_password>dbc</target_password> 
    <data_streams>1</data_streams> 
    <source_sessions>1</source_sessions> 
    <target_sessions>1</target_sessions> 
    <overwrite_existing_objects>TRUE</overwrite_existing_objects> 
    <freeze_job_steps>true</freeze_job_steps> 
    <force_utility>DSA</force_utility> 
    <log_level>1</log_level> 
    <online_archive>false</online_archive> 
    <database selection="unselected"> 
        <name>testdb</name> 
        <table selection="included"> 
            <name>fmt_inf</name> 
            <validate_row_count>ALL</validate_row_count> 
        </table> 
            <table selection="included"> 
                <name>test1</name> 
                <compare_ddl>true</compare_ddl> 
            </table> 
    </database> 
</dmCreate> 
In the next example, the XML file has been modified and is called changedParameters.xml. You could run start -job_name testStart -f changedParameters.xml - force_utility tptapi -sync to run a job that copies only table test3 and records where i = 2 from table fmt_inf to the target using TPTAPI.
<?xml version="1.0" encoding="UTF-8" standalone="yes" ?> 
<dmEdit xmlns="http://schemas.teradata.com/dataMover/v2009" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://schemas.teradata.com/unity/datamover.xsd">     
    <job_name>testStart</job_name> 
    <source_user>dmuser</source_user> 
    <source_password>dbc</source_password> 
    <target_user>dmuser</target_user> 
    <target_password>dbc</target_password> 
    <database selection="unselected"> 
        <name>testdb</name> 
        <table selection="included"> 
            <name>fmt_inf</name> 
            <validate_row_count>ALL</validate_row_count> 
            <compare_ddl>true</compare_ddl> 
            <sql_where_clause>
                <![CDATA[ where i = 2]]>
            </sql_where_clause> 
            <key_columns> 
                <key_column>i</key_column> 
            </key_columns> 
        </table> 
        <table selection="included"> 
            <name>test3</name> 
            <validate_row_count>ALL</validate_row_count> 
            <compare_ddl>true</compare_ddl> 
        </table> 
    </database> 
</dmEdit>