In this article, we are going to learn about the parameters of Import Oracle Datapump.
| Parameter | Description | 
| abort_step | undocumented feature | 
| access_method | data access method : default is automatic | 
| attach | attach to existing job : default is no | 
| cluster | start workers across cluster: default is y | 
| content | content to import : default is ALL | 
| data_options | Import data layer options | 
| current_edition | application on local database directory default directory specification | 
| directory | default directory specification | 
| dumper_directory | directory for stream dumper | 
| dumpfile | import dumpfile names format is (file1, file2…) | 
| encryption_password | encryption key to be used | 
| estimate | calculate size estimate default is BLOCKS | 
| exclude | import exclude option | 
| flashback_scn | system change number to be used for flashback import: default is no | 
| flashback_time | database time to be used for flashback import : default no | 
| full | indicates a full mode import | 
| help | help: display description of import parameters:default is N | 
| include | import include : default is no | 
| ip_address | ip address for PL/SQL debugger | 
| job_name | job_name : default is no | 
| keep master | keep master: retain job table upon completion | 
| logfile | log important messages to specified file | 
| master_only | only import the master table associated with this job | 
| metrics | enable/disable object metrics reporting | 
| mp_enable | enable/disable multi processing for current session | 
| network_link | network mod-import | 
| nologfile | no import log life creates | 
| package_load | specify how to load PL/SQL | 
| parallel | degree of parallelism default is 1 | 
| parallel_threshold | degree of DML parallelism | 
| parfile | parameter file: name of file contains parameter specification | 
| partition_options | options determine how partitions should be handle : default is NONE | 
| query | query used to select a subset of rows for a table | 
| remap_data | transform data is user tables | 
| remap_schema | remap source schema object to new schema | 
| remap_table | remap tables to a different name | 
| remap_tablespace | remap objects to a different name | 
| reuse_datafiles | re-initialize existing datafiles | 
| schemas | schemas to import: format is (‘schema1,….,schemaN’) | 
| services_name | service name that job will charge against | 
| silent | silent:display information default is NONE | 
| skip_unusable_indexes | skip indexes which indexes which are in the unused state | 
| source_edition | application edition to be used on remote database | 
| sqlfile | write appropiate sql DDL to speciied file | 
| status | interval between status updates | 
| streams_configuration | import streams configuration metadata | 
| table_exists_action | taken if the table to import already exists | 
| tables | tables to import : format is | 
| tablespaces | tablespaces to transport format is ‘(ts1,…, tsN)’ | 
| trace | trace option: enable sql_trace and timed_stat : default 0 | 
| transform | metadata_transforms | 
| transportable | use transportable data movement: default is NEVER | 
| transport_datafiles | list of datafiles to be plugged into target system | 
| transport_tablespaces | transportable tablespaces option: default is N | 
| transport_full_check | verify that tablespaces to be used do not have dependencies | 
| tts_closure_check | enable/disable transport able containment check: default is Y | 
| user_id | user/password to connect to oracle no default | 
| version | job version: compatible is default. | 
            