Saturday 13 June 2015

Parameter Mapping Original Export and Import to Datapump expdp and impdp

Parameter Mappings Between Old Export/Import and new Data Pump Expdp/Impdp:

This post describes how original Export and Import parameters map to the Data Pump Export and Import parameters that provides similar functionality.

Usage of Original Export Parameters with Data Pump Expdp

In following table will find how Data Pump Export accepts original Export parameters.  Table 1-1 describes how Data Pump Export interprets original Export parameters. Parameters that have the same name and functionality in both original Export and Data Pump Export are excluded from this table.
Table 1-1:Usage of Original Export Parameters with Data Pump Expdp


Original Export Parameter
 Data Pump Export Parameter
BUFFER
This parameter is ignored.
COMPRESS
Though we have COMPRESSION  parameter in Data Pump but it has different functionality from older Exp/Imp COMPRESS parameter.The old COMPRESS parameter is ignored. In original Export, the COMPRESS parameter affected how the initial extent was managed. Setting COMPRESS=n caused original Export to use current storage parameters for the initial and next extent.
The Data Pump Export COMPRESSION parameter is used to specify how data is compressed in the dump file, and is not related to the original Export COMPRESS parameter.
CONSISTENT
The old Exp/Imp CONSISTENT parameter is used to manage consistency of data in dump file.In Data Pump Export determines the current time and uses FLASHBACK_TIME parameter in place of older CONSISTENT parameter 
CONSTRAINTS
In old Export we have used CONSTRAINTS=n, and in Data Pump Export we use EXCLUDE=CONSTRAINTS.
The default behavior is to include constraints as part of the export.
DIRECT 
The  DIRECT parameter is ignored in new Data Pump Export. Data Pump Export automatically chooses the best export method.
FEEDBACK 
In older Export, feedback was given after a certain number of rows, as specified with the FEEDBACK command. In Data Pump Export, the status is given every mentioned seconds, as specified by STATUS.For example in the Data Pump Export STATUS=30 command is used. Note that this is not a direct mapping because the STATUS command returns the status of the export job, as well as the rows being processed.

FILE
In older export we use FILE parameter and in Data Pump Export we are using the DUMPFILE parameter, and also to determine whether a directory object exists to which the schema has read and write access.

GRANTS
In old Export we used GRANTS=n, and in Data Pump Export we use EXCLUDE=GRANT.
In old Export we used GRANTS=y, but in the Data Pump Export this parameter is ignored and does not need to be remapped because it is the Data Pump Export default behavior.
INDEXES
In old Export we used INDEXES=n, and in  Data Pump Export we use the EXCLUDE=INDEX parameter.
In old Export we used INDEXES=y, and in  Data Pump Export the parameter is ignored and does not need to be remapped because that is the Data Pump Export default behavior.
LOG
In place of older LOG parameter Data Pump Export attempts to determine the path that was specified or defaulted to for the LOGFILE parameter, and also to determine whether a directory object exists to which the schema has read and write access.

The contents of the log file will is about Data Pump Export operation.
OBJECT_CONSISTENT
This parameter is ignored because Data Pump Export processing ensures that each object is in a consistent state when being exported.
OWNER
In place of older OWNER parameter, the Data Pump SCHEMAS parameter is used.
RECORDLENGTH
This parameter is ignored because Data Pump Export automatically takes care of buffer sizing.
RESUMABLE
This parameter is ignored because Data Pump Export automatically provides this functionality to users who have been granted the EXP_FULL_DATABASE role.
RESUMABLE_NAME
This parameter is ignored because Data Pump Export automatically provides this functionality to users who have been granted the EXP_FULL_DATABASE role.
RESUMABLE_TIMEOUT
This parameter is ignored because Data Pump Export automatically provides this functionality to users who have been granted the EXP_FULL_DATABASE role.
ROWS
In place of old Export parameter  ROWS=y,  Data Pump Export uses the CONTENT=ALL parameter.
If place of old Export parameter ROWS=n, Data Pump Export uses the CONTENT=METADATA_ONLY parameter.
STATISTICS
This parameter is ignored because statistics are always saved for tables as part of a Data Pump export operation.
TABLESPACES
In old Export we used parameter TRANSPORT_TABLESPACE=n, in Data Pump Export  the TABLESPACES parameter is ignored.
In old Export we used parameter TRANSPORT_TABLESPACE=y, the Data Pump Export takes the names listed for the TABLESPACES parameter and uses them on the Data Pump Export TRANSPORT_TABLESPACES parameter.
TRANSPORT_TABLESPACE
If original Export used TRANSPORT_TABLESPACE=n (the default), then Data Pump Export uses the TABLESPACES parameter.
If original Export used TRANSPORT_TABLESPACE=y, then Data Pump Export uses the TRANSPORT_TABLESPACES parameter and only the metadata is exported.
TRIGGERS 
In old Export we used parameter  TRIGGERS=n, in Data Pump Export we are using  the EXCLUDE=TRIGGER parameter.
In old Export we used parameter  TRIGGERS=y,  the parameter is ignored in Data Pump Export and does not need to be remapped because that is the Data Pump Export default behavior.
TTS_FULL_CHECK 
In older Export we used TTS_FULL_CHECK=y, and in Data Pump Export we are using the TRANSPORT_FULL_CHECK parameter.
If original Export used TTS_FULL_CHECK=y, then the parameter is ignored and does not need to be remapped because that is the Data Pump Export default behavior.
VOLSIZE
When in older Export VOLSIZE parameter is used, it means the location specified for the dump file is a tape device. The Data Pump Export dump file format does not support tape devices. Therefore, this operation terminates with an error.

Usage of Old Import Parameters with Data Pump Impdp

In following table we will find how older Import parameters are mapped to new Data Pump Impdp parameters.
Table 1-2 :Usage of Old Import Parameters with Data Pump Impdp

Old Import Parameter
Data Pump Impdp Parameter
BUFFER
The BUFFER parameter is ignored.
CHARSET
The  parameter CHARSET was not supported several releases ago and should no longer be used. It will cause the Data Pump Import operation to abort.
COMMIT
The  COMMIT parameter is ignored. Data Pump Import automatically performs a commit after each table is processed.
COMPILE
The COMPILE  parameter is ignored. Data Pump Import compiles procedures after they are created. A recompile can be executed if necessary for dependency reasons.
CONSTRAINTS
In older Import we used CONSTRAINTS=n,and  in Data Pump Import we are using the EXCLUDE=CONSTRAINT parameter.
In older Import we used CONSTRAINTS=y, in Data Pump Import the parameter is ignored and does not need to be remapped because that is the Data Pump Import default behavior.
DATAFILES
In the Data Pump Import TRANSPORT_DATAFILES parameter is used.
DESTROY 
In older Import we used DESTROY=y, and  in Data Pump Import we are using the REUSE_DATAFILES=y parameter.
If older Import we used DESTROY=n, and in the Data Pump Import parameter is ignored and does not need to be remapped because that is the Data Pump Import default behavior.
FEEDBACK 
In old Import, feedback was given after a certain number of rows, as specified with the FEEDBACK command. In Data Pump Import, the status is given every mentioned seconds, as specified by STATUS.For example The Data Pump Import STATUS=30 command is used. Note that this is not a direct mapping because the STATUS command returns the status of the import job, as well as the rows being processed.

FILE
Data Pump Import uses the DUMPFILE parameter, and also to determine whether a directory object exists to which the schema has read and write access. In older import we used FILE parameter.

FILESIZE
The FILESIZE parameter is ignored because the information is already contained in the Data Pump dump file set.
FROMUSER
In the Data Pump Import SCHEMAS parameter is used. If FROMUSER was used without TOUSER also being used, then import schemas that have the IMP_FULL_DATABASE role cause Data Pump Import to attempt to create the schema and then import that schema's objects. Import schemas that do not have the IMP_FULL_DATABASE role can only import their own schema from the dump file set.
GRANTS
In old Import we used GRANTS=n,  and in Data Pump Import we are using  the EXCLUDE=OBJECT_GRANT parameter.
In old Import we used GRANTS=y, and in the Data Pump Import  this parameter is ignored and does not need to be remapped because that is the Data Pump Import default behavior.
IGNORE
In old Import we used IGNORE=y, and in Data Pump Import we are using the TABLE_EXISTS_ACTION=APPEND parameter. This causes the processing of table data to continue.
In old Import we used IGNORE=n, and in the Data Pump Import  this parameter is ignored and does not need to be remapped because that is the Data Pump Import default behavior.
INDEXES 
In old Import we used INDEXES=n, and in the Data Pump Import we are using the EXCLUDE=INDEX parameter.
In old Import we used used INDEXES=y, and in the Data Pump Import this parameter is ignored and does not need to be remapped because that is the Data Pump Import default behavior.
INDEXFILE
The Data Pump Import SQLFILE={directory-object:}filename and INCLUDE=INDEX parameters are used.
The same method and attempts made when looking for a directory object described for the FILE parameter also take place for the INDEXFILE parameter.
If no directory object was specified on the original Import, then Data Pump Import uses the directory object specified with the DIRECTORY parameter.
LOG
In Data Pump Import we use LOGFILE parameter, and also to determine whether a directory object exists to which the schema has read and write access.

The contents of the log file will be those of a Data Pump Import operation. 
RECORDLENGTH
This parameter is ignored because Data Pump handles issues about record length internally.
RESUMABLE
This parameter is ignored in Data Pump because this functionality is automatically provided for users who have been granted the IMP_FULL_DATABASE role.
RESUMABLE_NAME
This parameter is ignored in Data Pump because this functionality is automatically provided for users who have been granted the IMP_FULL_DATABASE role.
RESUMABLE_TIMEOUT
This parameter is ignored because this functionality is automatically provided for users who have been granted the IMP_FULL_DATABASE role.
ROWS=N 
In old Import we used ROWS=n, and in Data Pump Import we are using the CONTENT=METADATA_ONLY parameter.
In old Import we used ROWS=y, and in Data Pump Import we are using the CONTENT=ALL parameter.
SHOW
If in old import SHOW=y is specified, then the Data Pump Import SQLFILE=[directory_object:]file_name parameter is used to write the DDL for the import operation to a file. Only the DDL (not the entire contents of the dump file) is written to the specified file. (Note that the output is not shown on the screen as it was in original Import.)
The name of the file will be the file name specified on the DUMPFILEparameter (or on the original Import FILE parameter, which is remapped to DUMPFILE). If multiple dump file names are listed, then the first file name in the list is used. The file will be located in the directory object location specified on the DIRECTORY parameter or the directory object included on the DUMPFILE parameter. (Directory objects specified on theDUMPFILE parameter take precedence.)
STATISTICS
This parameter is ignored because statistics are always saved for tables as part of a Data Pump Import operation.
STREAMS_CONFIGURATION
This parameter is ignored because Data Pump Import automatically determines it; it does not need to be specified.
STREAMS_INSTANTIATION
This parameter is ignored because Data Pump Import automatically determines it; it does not need to be specified
TABLESPACES
In old Import we used TRANSPORT_TABLESPACE=n (the default),and in Data Pump Import  the TABLESPACES parameter is ignored.
In old Import we used TRANSPORT_TABLESPACE=y, and in Data Pump Import takes the names supplied for this TABLESPACES parameter and applies them to the Data Pump Import TRANSPORT_TABLESPACESparameter.
TOID_NOVALIDATE
This parameter is ignored. OIDs are no longer used for type validation.
TOUSER
The Data Pump Import REMAP_SCHEMA parameter is used. There may be more objects imported than with original Import. Also, Data Pump Import may create the target schema if it does not already exist.
The FROMUSER parameter must also have been specified in original Import or the operation will fail.
TRANSPORT_TABLESPACE
The TRANSPORT_TABLESPACE parameter is ignored, but if you also specified the DATAFILES parameter, then the import job continues to load the metadata. If the DATAFILES parameter is not specified, then an ORA-39002:invalid operation error message is returned.
TTS_OWNERS 
This parameter is ignored because this information is automatically stored in the Data Pump dump file set.
VOLSIZE
When the old Import VOLSIZE parameter is used, it means the location specified for the dump file is a tape device. The Data Pump Import dump file format does not support tape devices. Therefore, this operation terminates with an error.

2 comments: