Syntax and Description
If you specify a
directory_object, then it must be one that was previously established by the DBA and that you have access to. This overrides the directory object specified with the
DIRECTORY parameter. The default behavior is to create
log in the directory referenced by the directory object specified in the
file_name you specify already exists, then it will be overwritten.
All messages regarding work in progress, work completed, and errors encountered are written to the log file. (For a real-time status of the job, use the
STATUS command in interactive mode.)
A log file is always created unless the
NOLOGFILE parameter is specified. As with the dump file set, the log file is relative to the server and not the client.
Data Pump Import writes the log file using the database character set. If your client
NLS_LANG environment sets up a different client character set from the database character set, then it is possible that table names may be different in the log file than they are when displayed on the client output screen.
To perform a Data Pump Import using Oracle Automatic Storage Management (Oracle ASM), you must specify a
LOGFILE parameter that includes a directory object that does not include the Oracle ASM + notation. That is, the log file must be written to a disk file, and not written into the Oracle ASM storage. Alternatively, you can specify
NOLOGFILE=YES. However, this prevents the writing of the log file.
The following is an example of using the
LOGFILE parameter. You can create the
expfull.dmp dump file used in this example by running the example provided for the Export
FULL parameter. See "FULL".
> impdp hr SCHEMAS=HR DIRECTORY=dpump_dir2 LOGFILE=imp.log DUMPFILE=dpump_dir1:expfull.dmp
Because no directory object is specified on the
LOGFILE parameter, the log file is written to the directory object specified on the
"Using Directory Objects When Oracle Automatic Storage Management Is Enabled" for information about Oracle Automatic Storage Management and directory objects