(window.webpackJsonp=window.webpackJsonp||[]).push([[2338],{2746:function(a,t,s){"use strict";s.r(t);var e=s(31),r=Object(e.a)({},(function(){var a=this,t=a.$createElement,s=a._self._c||t;return s("ContentSlotsDistributor",{attrs:{"slot-key":a.$parent.slotKey}},[s("h1",{attrs:{id:"data-pump"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#data-pump"}},[a._v("#")]),a._v(" Data Pump")]),a._v(" "),s("p",[a._v("Following are the steps to create a data pump import/export:")]),a._v(" "),s("h2",{attrs:{id:"monitor-datapump-jobs"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#monitor-datapump-jobs"}},[a._v("#")]),a._v(" Monitor Datapump jobs")]),a._v(" "),s("p",[a._v("Datapump jobs can be monitored using")]),a._v(" "),s("p",[s("strong",[a._v("1. data dictionary views:")])]),a._v(" "),s("div",{staticClass:"language- extra-class"},[s("pre",{pre:!0,attrs:{class:"language-text"}},[s("code",[a._v("\n select * from dba_datapump_jobs; \n SELECT * FROM DBA_DATAPUMP_SESSIONS; \n select username,opname,target_desc,sofar,totalwork,message from V$SESSION_LONGOPS where username = 'bkpadmin';\n\n")])])]),s("p",[s("strong",[a._v("2. Datapump status:")])]),a._v(" "),s("ul",[s("li",[a._v("Note down the job name from the import/export logs or data dictionary name and")]),a._v(" "),s("li",[a._v("Run "),s("strong",[a._v("attach")]),a._v(" command:")]),a._v(" "),s("li",[a._v("type status in Import/Export prompt")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("impdp "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("bkpadmin"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("/")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("bkp123"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v(" attach"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("SYS_IMPORT_SCHEMA_01"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("Import")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("status")]),a._v("\n\n")])])]),s("p",[a._v("Press press "),s("strong",[a._v("CTRL+C")]),a._v(" to come out of Import/Export prompt")]),a._v(" "),s("h2",{attrs:{id:"step-3-6-create-directory"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#step-3-6-create-directory"}},[a._v("#")]),a._v(" Step 3/6 : Create directory")]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("create")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("or")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("replace")]),a._v(" directory DATAPUMP_REMOTE_DIR "),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("as")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token string"}},[a._v("'/oracle/scripts/expimp'")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(";")]),a._v("\n\n")])])]),s("h2",{attrs:{id:"step-7-export-commands"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#step-7-export-commands"}},[a._v("#")]),a._v(" Step 7 : Export Commands")]),a._v(" "),s("p",[a._v("Commands:")]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("expdp "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("bkpadmin"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("/")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("bkp123"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v(" parfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("exp"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("par"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\n\n")])])]),s("p",[a._v("*Please replace the data in <> with appropriate values as per your environment. You can add/modify parameters as per your requirements. In the above example all the remaining parameters are added in parameter files as stated below: *")]),a._v(" "),s("ul",[s("li",[a._v("Export Type : "),s("strong",[a._v("User Export")])]),a._v(" "),s("li",[a._v("Export entire schema")]),a._v(" "),s("li",[a._v("Parameter file details [say exp.par] :")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("schemas"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\ndirectory"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" DATAPUMP_REMOTE_DIR\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("dumpfile")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("dmp\nlogfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v("exp_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("log\n\n")])])]),s("ul",[s("li",[a._v("Export Type : "),s("strong",[a._v("User Export for large schema")])]),a._v(" "),s("li",[a._v("Export entire schema for large datasets: Here the export dump files will be broken down and compressed. Parallelism is used here "),s("strong",[a._v("(Note : Adding parallelism will increase the CPU load on server)")])]),a._v(" "),s("li",[a._v("Parameter file details [say exp.par] :")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("schemas"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\ndirectory"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" DATAPUMP_REMOTE_DIR\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("dumpfile")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("%")]),a._v("U"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("dmp\nlogfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v("exp_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("log \ncompression "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("all")]),a._v("\nparallel"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token number"}},[a._v("5")]),a._v("\n\n")])])]),s("ul",[s("li",[a._v("Export Type : "),s("strong",[a._v("Table Export")]),a._v(" [ Export set of tables]")]),a._v(" "),s("li",[a._v("Parameter file details [say exp.par] :")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("tables")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" tname1"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(",")]),a._v(" tname2"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(",")]),a._v(" tname3\ndirectory"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" DATAPUMP_REMOTE_DIR\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("dumpfile")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("dmp\nlogfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v("exp_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("log\n\n")])])]),s("h2",{attrs:{id:"step-9-import-commands"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#step-9-import-commands"}},[a._v("#")]),a._v(" Step 9 : Import Commands")]),a._v(" "),s("p",[a._v("Prerequisite:")]),a._v(" "),s("ul",[s("li",[a._v("Prior to user import it is a good practice to drop the schema or table imported.")])]),a._v(" "),s("p",[a._v("Commands:")]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("impdp "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("bkpadmin"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("/")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("bkp123"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v(" parfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("imp"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("par"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\n\n")])])]),s("p",[a._v("*Please replace the data in <> with appropriate values as per your environment. You can add/modify parameters as per your requirements. In the above example all the remaining parameters are added in parameter files as stated below: *")]),a._v(" "),s("ul",[s("li",[a._v("Import Type : "),s("strong",[a._v("User Import")])]),a._v(" "),s("li",[a._v("Import entire schema")]),a._v(" "),s("li",[a._v("Parameter file details [say imp.par] :")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("schemas"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\ndirectory"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" DATAPUMP_REMOTE_DIR\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("dumpfile")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("dmp\nlogfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v("imp_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("log\n\n")])])]),s("ul",[s("li",[a._v("Import Type : "),s("strong",[a._v("User Import for large schema")])]),a._v(" "),s("li",[a._v("Import entire schema for large datasets: Parallelism is used here "),s("strong",[a._v("(Note : Adding parallelism will increase the CPU load on server)")])]),a._v(" "),s("li",[a._v("Parameter file details [say imp.par] :")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[a._v("schemas"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\ndirectory"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" DATAPUMP_REMOTE_DIR\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("dumpfile")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("%")]),a._v("U"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("dmp\nlogfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v("imp_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("log \nparallel"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token number"}},[a._v("5")]),a._v("\n\n")])])]),s("ul",[s("li",[a._v("Import Type : "),s("strong",[a._v("Table Import")]),a._v(" [ Import set of tables]")]),a._v(" "),s("li",[a._v("Parameter file details [say imp.par] :")])]),a._v(" "),s("div",{staticClass:"language-sql extra-class"},[s("pre",{pre:!0,attrs:{class:"language-sql"}},[s("code",[s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("tables")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" tname1"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(",")]),a._v(" tname2"),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(",")]),a._v(" tname3\ndirectory"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" DATAPUMP_REMOTE_DIR\n"),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("dumpfile")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("dmp\nlogfile"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v("exp_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("dbname"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("_"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("schema")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),s("span",{pre:!0,attrs:{class:"token punctuation"}},[a._v(".")]),a._v("log\nTABLE_EXISTS_ACTION"),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("=")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("<")]),a._v("APPEND "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("/")]),a._v("SKIP "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("/")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("TRUNCATE")]),a._v(" "),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v("/")]),s("span",{pre:!0,attrs:{class:"token keyword"}},[a._v("REPLACE")]),s("span",{pre:!0,attrs:{class:"token operator"}},[a._v(">")]),a._v("\n\n")])])]),s("h2",{attrs:{id:"_1-datapump-steps"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#_1-datapump-steps"}},[a._v("#")]),a._v(" 1. Datapump steps")]),a._v(" "),s("table",[s("thead",[s("tr",[s("th",[a._v("Source Server [Export Data]")]),a._v(" "),s("th",[a._v("Target Server [Import Data]")])])]),a._v(" "),s("tbody",[s("tr",[s("td",[a._v("1. Create a datapump folder that will contain the export dump files")]),a._v(" "),s("td",[a._v("4. Create a datapump folder that will contain the import dump files")])]),a._v(" "),s("tr",[s("td",[a._v("2. Login to database schema that will perform the export.")]),a._v(" "),s("td",[a._v("5. Login to database schema that will perform the import.")])]),a._v(" "),s("tr",[s("td",[a._v("3. Create directory pointing to step 1.")]),a._v(" "),s("td",[a._v("6. Create directory pointing to step 4.")])]),a._v(" "),s("tr",[s("td",[a._v("7. Run Export Statements.")]),a._v(" "),s("td")]),a._v(" "),s("tr",[s("td",[a._v("8. Copy/SCP the dump files to Target Server.")]),a._v(" "),s("td")]),a._v(" "),s("tr",[s("td"),a._v(" "),s("td",[a._v("9. Run Import statements")])]),a._v(" "),s("tr",[s("td"),a._v(" "),s("td",[a._v("10. check data ,compile invalid objects and provide related grants")])])])]),a._v(" "),s("h2",{attrs:{id:"copy-tables-between-different-schemas-and-tablespaces"}},[s("a",{staticClass:"header-anchor",attrs:{href:"#copy-tables-between-different-schemas-and-tablespaces"}},[a._v("#")]),a._v(" Copy tables between different schemas and tablespaces")]),a._v(" "),s("div",{staticClass:"language- extra-class"},[s("pre",{pre:!0,attrs:{class:"language-text"}},[s("code",[a._v("\nexpdp / directory=DATAPUMP_REMOTE_DIR dumpfile=\n\n\n impdp / directory=DATAPUMP_REMOTE_DIR dumpfile= remap_schema=: remap_tablespace=:\n\n")])])])])}),[],!1,null,null,null);t.default=r.exports}}]);