In this article, we are going to learn about the parameters of Import Oracle Datapump.

Parameter Description
abort_step undocumented feature
access_method data access method : default is automatic
attach attach to existing job : default is no
cluster start workers across cluster: default is y
content content to import : default is ALL
data_options Import data layer options
current_edition application on local database directory default directory specification
directory default directory specification
dumper_directory directory for stream dumper
dumpfile import dumpfile names format is (file1, file2…)
encryption_password encryption key to be used
estimate calculate size estimate default is BLOCKS
exclude import exclude option
flashback_scn system change number to be used for flashback import: default is no
flashback_time database time to be used for flashback import : default no
full indicates a full mode import
help help: display description of import parameters:default is N
include import include : default is no
ip_address ip address for PL/SQL debugger
job_name job_name : default is no
keep master keep master: retain job table upon completion
logfile log important messages to specified file
master_only only import the master table associated with this job
metrics enable/disable object metrics reporting
mp_enable enable/disable multi processing for current session
network_link network mod-import
nologfile no import log life creates
package_load specify how to load PL/SQL
parallel degree of parallelism default is 1
parallel_threshold degree of DML parallelism
parfile parameter file: name of file contains parameter specification
partition_options options  determine how partitions should be handle : default is NONE
query query used to select a subset of rows for a table
remap_data transform data is user tables
remap_schema remap source schema object to new schema
remap_table remap tables to a different name
remap_tablespace remap objects to a different name
reuse_datafiles re-initialize existing datafiles
schemas schemas to import: format is (‘schema1,….,schemaN’)
services_name service name that job will charge against
silent silent:display  information default is NONE
skip_unusable_indexes skip indexes which indexes which are in the unused state
source_edition application edition to be used  on remote database
sqlfile write appropiate sql DDL to speciied file
status interval between status updates
streams_configuration import streams configuration metadata
table_exists_action taken if  the table to import already exists
tables tables to import : format is
tablespaces tablespaces to transport format is  ‘(ts1,…, tsN)’
trace trace option: enable sql_trace and timed_stat : default 0
transform metadata_transforms
transportable use transportable data movement: default is NEVER
transport_datafiles list of datafiles to be plugged into target system
transport_tablespaces transportable tablespaces option: default is N
transport_full_check verify that tablespaces to be  used do not have dependencies
tts_closure_check enable/disable transport able containment check: default is Y
user_id user/password to connect to oracle no default
version job version: compatible is default.



About The Author

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.