[2016-6-NEW]Free Braindump2go 1Z0-060 Exam Dumps Free 161Q&As Download[NQ1-NQ10]

2016 June Oracle Official 1Z0-060: Upgrade to Oracle Database 12c Exam Questions New Updated Today in Braindump2go.com. 100% 1Z0-060 Exam Pass Guaranteed!


You notice that the performance of your production 24/7 Oracle database significantly degraded. Sometimes you are not able to connect to the instance because it hangs.
You do not want to restart the database instance.
How can you detect the cause of the degraded performance?

A.    Enable Memory Access Mode, which reads performance data from SGA.
B.    Use emergency monitoring to fetch data directly from SGA analysis.
C.    Run Automatic Database Diagnostic Monitor (ADDM) to fetch information from the latest Automatic
Workload Repository (AWR) snapshots.
D.    Use Active Session History (ASH) data and hang analysis in regular performance monitoring.
E.    Run ADDM in diagnostic mode.

Answer: B

You plan to use the In Database Archiving feature of Oracle Database 12c, and store rows that are inactive for over three months, in Hybrid Columnar Compressed (HCC) format.
Which three storage options support the use of HCC?

A.    ASM disk groups with ASM disks consisting of Exadata Grid Disks.
B.    ASM disk groups with ASM disks consisting of LUNS on any Storage Area Network array
C.    ASM disk groups with ASM disks consisting of any zero padded NFS-mounted files
D.    Database files stored in ZFS and accessed using conventional NFS mounts.
E.    Database files stored in ZFS and accessed using the Oracle Direct NFS feature
F.    Database files stored in any file system and accessed using the Oracle Direct NFS feature
G.    ASM disk groups with ASM disks consisting of LUNs on Pillar Axiom Storage arrays

Answer: AEG
HCC requires the use of Oracle Storage Exadata(A), Pillar Axiom(G)or Sun ZFS Storage Appliance (ZFSSA).
*Hybrid Columnar Compression, initially only available on Exadata, has been extended to supportPillar Axiom and Sun ZFS Storage Appliance (ZFSSA) storage when used with Oracle Database Enterprise Edition and above
*Oracle offers the ability to manage NFS using a feature called Oracle Direct NFS (dNFS). Oracle Direct NFS implements NFS V3 protocol within the Oracle database kernel itself. Oracle Direct NFS client overcomes many of the challenges associated with using NFS with the Oracle Database with simple configuration, betterperformance than traditional NFS clients, and offers consistent configuration across platforms.

In your multitenant container database (CDB) containing pluggable databases (PDB), users complain about performance degradation.
How does real-time Automatic database Diagnostic Monitor (ADDM) check performance degradation and provide solutions?

A.    It collects data from SGA and compares it with a preserved snapshot.
B.    It collects data from SGA, analyzes it, and provides a report.
C.    It collects data from SGA and compares it with the latest snapshot.
D.    It collects data from both SGA and PGA, analyzes it, and provides a report.

Answer: B
*The multitenant architecture enables an Oracle database to function as a multitenant container database (CDB) that includes zero, one, or many customer-created pluggable databases (PDBs). A PDB is a portable collection of schemas, schema objects, and nonschema objects that appears to an Oracle Net client as a non-CDB. All Oracle databases before Oracle Database 12c were non-CDBs.
*The System Global Area (SGA) is a group of shared memory areas that are dedicated to an Oracle “instance” (an instance is your database programs and RAM).
*The PGA (Program or Process Global Area) is a memory area (RAM) that stores data and control information for a single process.

The tnsnames.ora file has an entry for the service alias ORCL as follows:
The TNS ping command executes successfully when tested with ORCL; however, from the same OS user session, you are not able to connect to the database instance with the following command:
SQL > CONNECT scott/tiger@orcl
What could be the reason for this?

A.    The listener is not running on the database node.
B.    The TNS_ADMIN environment variable is set to the wrong value.
C.    The orcl.oracle.com database service is not registered with the listener.
D.    The DEFAULT_DOMAIN parameter is set to the wrong value in the sqlnet.ora file.
E.    The listener is running on a different port.

Answer: C
Service registration enables the listener to determine whether a database service and its service handlers are available. A service handler is a dedicated server process or dispatcher that acts as a connection point to a database. During registration, the LREG process provides the listener with the instance name, database service names, and the type and addresses of service handlers. This information enables the listener to start a service handler when a client request arrives.

Examine the following steps of privilege analysis for checking and revoking excessive, unused privileges granted to users:
1. Create a policy to capture the privilege used by a user for privilege analysis.
2. Generate a report with the data captured for a specified privilege capture.
3. Start analyzing the data captured by the policy.
4. Revoke the unused privileges.
5. Compare the used and unused privileges’ lists.
6. Stop analyzing the data.
Identify the correct sequence of steps.

A.    1, 3, 5, 6, 2, 4
B.    1, 3, 6, 2, 5, 4
C.    1, 3, 2, 5, 6, 4
D.    1, 3, 2, 5, 6, 4
E.    1, 3, 5, 2, 6, 4

Answer: B
Define the privilege analysis policy.
Enable the privilege analysis policy.
This step begins recording the privilege use that the policy defined.
Disable the privilege analysis policy’s recording of privilege use.
This step stops capturing the privilege use for the policy.
Generate privilege analysis results.
This step writes the results to the data dictionary views described in “Privilege Analysis Policy and Report Data Dictionary Views”.
Optionally, disable and then drop the privilege analysis policy.
Dropping a privilege analysis policy deletes the data captured by the policy.

You database is running an ARCHIVELOG mode.
The following parameter are set in your database instance:
LOG_ARCHIVE_FORMAT = arch+%t_%r.arc
LOG_ARCHIVE_DEST_1 = `LOCATION = /disk1/archive’
DB_RECOVERY_FILE = `/u01/oradata’
Which statement is true about the archived redo log files?

A.    They are created only in the location specified by the LOG_ARCHIVE_DEST_1 parameter.
B.    They are created only in the Fast Recovery Area.
C.    They are created in the location specified by the LOG_ARCHIVE_DEST_1 parameter and in the
default location $ORACLE_HOME/dbs/arch.
D.    They are created in the location specified by the LOG_ARCHIVE_DEST_1 parameter and the location
specified by the DB_RECOVERY_FILE_DEST parameter.

Answer: A
You can choose to archive redo logs to a single destination or to multiple destinations.
Destinations can be local–within the local file system or an Oracle Automatic Storage Management (Oracle ASM) disk group–or remote (on a standby database). When you archive to multiple destinations, a copy of each filled redo log file is written to each destination. These redundant copies help ensure that archived logs are always available in the event of a failure at one of the destinations. To archive to only a single destination, specify that destination using the LOG_ARCHIVE_DESTandLOG_ARCHIVE_DUPLEX_DESTinitialization parameters. ARCHIVE_DEST initialization parameter. To archive to multiple destinations, you can choose to archive to two or more locations using the LOG_ARCHIVE_DEST_n initialization parameters, or to archive only to a primary and secondary destination using the LOG_ ARCHIVE_DEST and LOG_ARCHIVE_DUPLEX_DEST initialization parameters.

Your multitenant container database (CDB) is running in ARCHIVELOG mode. You connect to the CDB RMAN.
Examine the following command and its output:
You execute the following command:
Which data files will be backed up?

A.    Data files that belong to only the root container
B.    Data files that belong to the root container and all the pluggable databases (PDBs)
C.    Data files that belong to only the root container and PDB$SEED
D.    Data files that belong to the root container and all the PDBs excluding PDB$SEED

Answer: B
Backing Up a Whole CDB
Backing up a whole CDB is similar to backing up a non-CDB. When you back up a whole CDB, RMAN backs up the root, all the PDBs, and the archived redo logs.
You can then recover either the whole CDB, the root only, or one or more PDBs from the CDB backup.
*You can back up and recover a whole CDB, the root only, or one or more PDBs.
*Backing Up Archived Redo Logs with RMAN
Archived redo logs are the key to successful media recovery. Back them up regularly. You can back up logs with BACKUP ARCHIVELOG, or back up logs while backing up datafiles and control files by specifying BACKUP … PLUS ARCHIVELOG.

You are administering a database stored in Automatic Storage management (ASM). The files are stored in the DATA disk group. You execute the following command:
SQL > ALTER DISKGROUP data ADD ALIAS `+data/prod/myfile.dbf’ FOR `+data.231.45678′;
What is the result?

A.    The file `+data.231.45678′ is physically relocated to `+data/prod’ and renamed as `myfile.dbf’.
B.    The file `+data.231.45678′ is renamed as `myfile.dbf’, and copied to `+data/prod’.
C.    The file `+data.231.45678′ remains in the same location andasynonym’myfile.dbf’ is created.
D.    The file `myfile.dbf’ is created in `+data/prod’ and the reference to `+data.231.45678′ in the data
dictionary removed.

Answer: C
Use this clause to create an alias name for an Oracle ASM filename. The alias_name consists of the full directory path and the alias itself.

Which three functions are performed by the SQL Tuning Advisor?

A.    Building and implementing SQL profiles
B.    Recommending the optimization of materialized views
C.    Checking query objects for missing and stale statistics
D.    Recommending bitmap, function-based, and B-tree indexes
E.    Recommending the restructuring of SQL queries that are using bad plans

Answer: ACE
The SQL Tuning Advisor takes one or more SQL statements as an input and invokes the Automatic Tuning Optimizer to perform SQL tuning on the statements. The output of the SQL Tuning Advisor is in the form of an advice or recommendations, along with a rationale for eachrecommendation and its expected benefit. The recommendation relates to collection of statistics on objects(C), creation of new indexes, restructuring of the SQL statement(E), or creation of a SQL profile(A). You can choose to accept the recommendation to complete the tuning of the SQL statements.

Examine the following command:
ALTER SYSTEM SET enable_ddl_logging=FALSE;
Which statement is true?

A.    None of the data definition language (DDL) statements are logged in the trace file.
B.    Only DDL commands that resulted in errors are logged in the alert log file.
C.    A new log.xml file that contains the DDL statements is created, and the DDL command details are
removed from the alert log file.
D.    Only DDL commands that resulted in the creation of new database files are logged.

Answer: A
ENABLE_DDL_LOGGING enables or disables the writing of a subset of data definition language (DDL) statements to a DDL alert log.
The DDL log is a file that has the same format and basic behavior as the alert log, but it only contains the DDL statements issued by the database. The DDL log is created only for the RDBMS component and only if the ENABLE_DDL_LOGGING initialization parameter is set to true. When this parameter is set to false, DDL statements are not included in any log.

2016 Valid Braindump2go Oracle 1Z0-060 Study Materials:


1.| Braindump2go Latest 1Z0-060 PDF and 1Z0-060 VCE Dumps with New 161q 1Z0-060 Exam Questions: http://www.braindump2go.com/1z0-060.html

[100% Exam Pass Guaranteed!]


2.| Braindump2go New 1Z0-060 Exam Questions and Answers – Google Drive: https://drive.google.com/folderview?id=0B75b5xYLjSSNWm1lN2tlY1ZKVVk&usp=sharing


MORE Practice is the Most Important IF You want to PASS Oracle 1Z0-060 Exam 100%!
————— Braindump2go.com
————— Pass All IT Exams at the first Try!

Comments are closed.