We have professional as well as dedicated This experts that have rich experience in compiling the actual Oracle Oracle exam braindumps. They will instruct that you well put together for the Oracle 1Z0-062 exam. And you will also visit each of our interactive forum at Examcollection house page. You can find many useful tips as well as suggestions there. Simply because a great amount of people possess passed the actual exam after employing our Oracle 1Z0-062 practice demos. They come again to Examcollection as well as share their own experience along with all the candidates. Almost all of them speak extremely of each of our Oracle Oracle 1Z0-062 practice questions and answers. Your precious suggestions may benefit you a lot.
2021 Dec 1Z0-062 actual test
Q71. You enabled an audit policy by issuing the following statements:
SQL> AUDIT POLICY ORA_DATABASE_PARAMETER BY SCOTT;
SQL> AUDIT POLICY ORA_DATABASE_PARAMETER BY SYS, SYSTEM;
For which database users and for which executions is the audit policy now active? Select two.
A. SYS, SYSTEM
C. Only for successful executions
D. Only for failed executions
E. Both successful and failed executions
Explanation: * The ORA_DATABASE_PARAMETER policy audits commonly used Oracle Database parameter settings. By default, this policy is not enabled.
Q72. Which two statements are true about the Automatic Database Diagnostic Monitor (ADDM)? (Choose two.)
A. The ADDM requires at least four AWR snapshots for analysis
B. The ADDM runs after each AWR snapshot is collected automatically by MMON
C. The results of the ADDM analysis are stored in the Automatic Workload Repository (AWR)
D. The ADDM analysis provides only diagnostics information but does not provide recommendations
E. The ADDM calls other advisors if required, but does not provide recommendations about the advisors
Q73. Examine the commands executed to monitor database operations:
$> conn sys oracle/oracle@prod as sysdba
SQL > VAR eid NUMBER
SQL > EXEC: eid := DBMS_SQL_MONITOR.BEGIN_OPERATION (‘batch_job’ , FORCED_TRACKING => ‘Y’);
Which two statements are true?
A. Database operations will be monitored only when they consume a significant amount of resource.
B. Database operations for all sessions will be monitored.
C. Database operations will be monitored only if the STATISTICS_LEVEL parameter is set to TYPICAL and CONTROL_MANAGEMENT_PACK_ACCESS is set DIAGNISTIC + TUNING.
D. Only DML and DDL statements will be monitored for the session.
E. All subsequent statements in the session will be treated as one database operation and will be monitored.
Explanation: C: Setting the CONTROL_MANAGEMENT_PACK_ACCESS initialization parameter to DIAGNOSTIC+TUNING (default) enables monitoring of database operations. Real-Time SQL Monitoring is a feature of the Oracle Database Tuning Pack.
* The DBMS_SQL_MONITOR package provides information about Real-time SQL
Monitoring and Real-time Database Operation Monitoring.
*(not B) BEGIN_OPERATION Function
starts a composite database operation in the current session.
/ (E) FORCE_TRACKING - forces the composite database operation to be tracked when the operation starts. You can also use the string variable 'Y'.
/ (not A) NO_FORCE_TRACKING - the operation will be tracked only when it has consumed at least 5 seconds of CPU or I/O time. You can also use the string variable 'N'.
Q74. Your are the DBA supporting an Oracle 11g Release 2 database and wish to move a table containing several DATE, CHAR, VARCHAR2, and NUMBER data types, and the table’s indexes, to another tablespace.
The table does not have a primary key and is used by an OLTP application.
Which technique will move the table and indexes while maintaining the highest level of availability to the application?
A. Oracle Data Pump.
B. An ALTER TABLE MOVE to move the table and ALTER INDEX REBUILD to move the indexes.
C. An ALTER TABLE MOVE to move the table and ALTER INDEX REBUILD ONLINE to move the indexes.
D. Online Table Redefinition.
E. Edition-Based Table Redefinition.
Explanation: * Oracle Database provides a mechanism to make table structure modifications without significantly affecting the availability of the table. The mechanism is called online table redefinition. Redefining tables online provides a substantial increase in availability compared to traditional methods of redefining tables.
* To redefine a table online:
Choose the redefinition method: by key or by rowid
* By key—Select a primary key or pseudo-primary key to use for the redefinition. Pseudo-primary keys are unique keys with all component columns having NOT NULL constraints. For this method, the versions of the tables before and after redefinition should have the same primary key columns. This is the preferred and default method of redefinition.
* By rowid—Use this method if no key is available. In this method, a hidden column named M_ROW$$ is added to the post-redefined version of the table. It is recommended that this column be dropped or marked as unused after the redefinition is complete. If COMPATIBLE is set to 10.2.0 or higher, the final phase of redefinition automatically sets this column unused. You can then use the ALTER TABLE ... DROP UNUSED COLUMNS statement to drop it.
You cannot use this method on index-organized tables.
* When you rebuild an index, you use an existing index as the data source. Creating an index in this manner enables you to change storage characteristics or move to a new tablespace. Rebuilding an index based on an existing data source removes intra-block fragmentation. Compared to dropping the index and using the CREATE INDEX statement, re-creating an existing index offers better performance.
Not E: Edition-based redefinition enables you to upgrade the database component of an application while it is in use, thereby minimizing or eliminating down time.
Q75. You use a recovery catalog for maintaining your database backups.
You execute the following command: $rman TARGET / CATALOG rman / cat@catdb RMAN > BACKUP VALIDATE DATABASE ARCHIVELOG ALL;
Which two statements are true?
A. Corrupted blocks, if any, are repaired.
B. Checks are performed for physical corruptions.
C. Checks are performed for logical corruptions.
D. Checks are performed to confirm whether all database files exist in correct locations
E. Backup sets containing both data files and archive logs are created.
Explanation: B (not C): You can validate that all database files and archived redo logs can be backed up by running a command as follows:
RMAN> BACKUP VALIDATE DATABASE ARCHIVELOG ALL;
This form of the command would check for physical corruption. To check for logical corruption,
RMAN> BACKUP VALIDATE CHECK LOGICAL DATABASE ARCHIVELOG ALL;
D: You can use the VALIDATE keyword of the BACKUP command to do the following:
Check datafiles for physical and logical corruption
Confirm that all database files exist and are in the correct locations.
Note: You can use the VALIDATE option of the BACKUP command to verify that database files exist and are in the correct locations (D), and have no physical or logical corruptions that would prevent RMAN from creating backups of them. When performing a BACKUP...VALIDATE, RMAN reads the files to be backed up in their entirety, as it would during a real backup. It does not, however, actually produce any backup sets or image copies (Not A, not E).
Renew 1Z0-062 test preparation:
Q76. Which two are prerequisites for performing a flashback transaction?
A. Flashback Database must be enabled.
B. Undo retention guarantee for the database must be configured.
C. EXECUTE privilege on the DBMS_FLASHBACK package must be granted to the user flashing back transaction.
D. Supplemental logging must be enabled.
E. Recycle bin must be enabled for the database.
F. Block change tracking must be enabled tor the database.
Reference: Oracle Database Advanced Application Developer's Guide 11g, Using Oracle Flashback Technology
Q77. Your database has the SRV1 service configured for an application that runs on middle-tier application server. The application has multiple modules. You enable tracing at the service level by executing the following command:
SQL > exec DBMS_MONITOR.SERV_MOD_ACT_TRACE_ENABLE (‘SRV1’);
The possible outcome and actions to aggregate the trace files are as follows:
1. The command fails because a module name is not specified.
2. A trace file is created for each session that is running the SRV1 service.
3. An aggregated trace file is created for all the sessions that are running the SRV1 service.
4. The trace files may be aggregated by using the trcess utility.
5. The trace files be aggregated by using the tkprof utility.
Identify the correct outcome and the step to aggregate by using tkprof utility?
B. 2 and 4
C. 2 and 5
D. 3 and 4
E. 3 and 5
Explanation: Tracing information is present in multiple trace files and you must use the trcsess tool to collect it into a single file.
Not 1: Parameter service_name
Name of the service for which tracing is enabled.
Name of the MODULE. An optional additional qualifier for the service.
* The procedure enables a trace for a given combination of Service, MODULE and ACTION name. The specification is strictly hierarchical: Service Name or Service Name/MODULE, or Service Name, MODULE, and ACTION name must be specified. Omitting a qualifier behaves like a wild-card, so that not specifying an ACTION means all ACTIONs. Using the ALL_ACTIONS constant achieves the same purpose.
* SERV_MOD_ACT_TRACE_ENABLE Procedure
This procedure will enable SQL tracing for a given combination of Service Name, MODULE and ACTION globally unless an instance_name is specified.
service_name IN VARCHAR2,
module_name IN VARCHAR2 DEFAULT ANY_MODULE,
action_name IN VARCHAR2 DEFAULT ANY_ACTION,
waits IN BOOLEAN DEFAULT TRUE,
binds IN BOOLEAN DEFAULT FALSE,
instance_name IN VARCHAR2 DEFAULT NULL);
Q78. In order to exploit some new storage tiers that have been provisioned by a storage administrator, the partitions of a large heap table must be moved to other tablespaces in your Oracle 12c database?
Both local and global partitioned B-tree Indexes are defined on the table.
A high volume of transactions access the table during the day and a medium volume of transactions access it at night and during weekends.
Minimal disrupt ion to availability is required.
Which three statements are true about this requirement?
A. The partitions can be moved online to new tablespaces.
B. Global indexes must be rebuilt manually after moving the partitions.
C. The partitions can be compressed in the same tablespaces.
D. The partitions can be compressed in the new tablespaces.
E. Local indexes must be rebuilt manually after moving the partitions.
Explanation: A: You can create and rebuild indexes online. Therefore, you can update base tables at the same time you are building or rebuilding indexes on that table. You can perform DML operations while the index build is taking place, but DDL operations are not allowed. Parallel execution is not supported when creating or rebuilding an index online.
D: Moving (Rebuilding) Index-Organized Tables Because index-organized tables are primarily stored in a B-tree index, you can encounter fragmentation as a consequence of incremental updates. However, you can use the ALTER TABLE...MOVE statement to rebuild the index and reduce this fragmentation.
C: If a table can be compressed in the new tablespace, also it can be compressed in the same tablespace.
Not B, not E: Local and Global indexes can be automatically rebuild with UPDATE
INDEXES when you move the table.
Q79. Which two are true concerning a multitenant container database with three pluggable database?
A. All administration tasks must be done to a specific pluggable database.
B. The pluggable databases increase patching time.
C. The pluggable databases reduce administration effort.
D. The pluggable databases are patched together.
E. Pluggable databases are only used for database consolidation.
Explanation: The benefits of Oracle Multitenant are brought by implementing a pure deployment choice. The following list calls out the most compelling examples.
* High consolidation density. (E) The many pluggable databases in a single multitenant container database share its memory and background processes, letting you operate many more pluggable databases on a particular platform than you can single databases that use the old architecture. This is the same benefit that schema-based consolidation brings.
* Rapid provisioning and cloning using SQL.
* New paradigms for rapid patching and upgrades. (D, not B) The investment of time and effort to patch one multitenant container database results in patching all of its many pluggable databases. To patch a single pluggable database, you simply unplug/plug to a multitenant container database at a different Oracle Database software version.
* (C, not A) Manage many databases as one. By consolidating existing databases as pluggable databases, administrators can manage many databases as one. For example, tasks like backup and disaster recovery are performed at the multitenant container database level.
* Dynamic between pluggable database resource management. In Oracle Database 12c, Resource Manager is extended with specific functionality to control the competition for resources between the pluggable databases within a multitenant container database. Note:
* Oracle Multitenant is a new option for Oracle Database 12c Enterprise Edition that helps customers reduce IT costs by simplifying consolidation, provisioning, upgrades, and more. It is supported by a new architecture that allows a multitenant container database to hold many pluggable databases. And it fully complements other options, including Oracle Real Application Clusters and Oracle Active Data Guard. An existing database can be simply adopted, with no change, as a pluggable database; and no changes are needed in the other tiers of the application.
Reference: 12c Oracle Multitenant
Q80. You upgraded from a previous Oracle database version to Oracle Database version to Oracle Database 12c. Your database supports a mixed workload. During the day, lots of insert, update, and delete operations are performed. At night, Extract, Transform, Load (ETL) and batch reporting jobs are run. The ETL jobs perform certain database operations using two or more concurrent sessions.
After the upgrade, you notice that the performance of ETL jobs has degraded. To ascertain the cause of performance degradation, you want to collect basic statistics such as the level of parallelism, total database time, and the number of I/O requests for the ETL jobs.
How do you accomplish this?
A. Examine the Active Session History (ASH) reports for the time period of the ETL or batch reporting runs.
B. Enable SQL tracing for the queries in the ETL and batch reporting queries and gather diagnostic data from the trace file.
C. Enable real-time SQL monitoring for ETL jobs and gather diagnostic data from the V$SQL_MONITOR view.
D. Enable real-time database operation monitoring using the DBMS_SQL_MONITOR.BEGIN_OPERATION function, and then use the DBMS_SQL_MONITOR.REPORT_SQL_MONITOR function to view the required information.
Explanation: * Monitoring database operations Real-Time Database Operations Monitoring enables you to monitor long running database tasks such as batch jobs, scheduler jobs, and Extraction, Transformation, and Loading (ETL) jobs as a composite business operation. This feature tracks the progress of SQL and PL/SQL queries associated with the business operation being monitored. As a DBA or developer, you can define business operations for monitoring by explicitly specifying the start and end of the operation or implicitly with tags that identify the operation.