Cisco Cisco Workload Automation 6.3 User Guide

Page of 46
29
 
Using the Sqoop Adapter
Defining Sqoop Jobs
Export Job Task
Details Tab
Note: 
You can click the Variables button to insert a predefined variable into a selected field on this tab.
Source Path
 – HDFS source path for the export that contains the source data.
Destination Table
 – Select the table you want to use to populate in the database.
Existing Jar File 
– Enter the name of the jar to export the record class from.
Allow Updates
 – Depending on the target database, you can update rows if they exist in the database already or 
insert rows if they do not exist yet. By default, the export is carried out using INSERT statements. By checking the 
Allow Updates
 option, you can choose to use UPDATE statements or INSERT and UPDATE statements together. 
The list box next to the 
Allow Updates
 option allows you to select the desired update mode.
Key Columns for Updates
 – Anchor columns to use for updates. Use a comma separated list of columns if there 
is more than one column. This is a mandatory field if 
Allow Updates 
is selected.
Use Batch Mode
 – Use batch mode for underlying statement execution.
Use High Performance Direct Export
 – Select to use the direct export fast path. 
MySQL provides a direct mode for exports as well, using the 
mysqlimport
 tool. This may be higher-performance 
than the standard JDBC codepath. When using export in direct mode with MySQL, the MySQL bulk utility 
mysqlimport
 must be available in the shell path of the task process.
Number of Mappers
 – Select the number of map tasks to export in parallel.