web analytics

Braindump2go Free Microsoft Dumps Questions Collection

Latest Real Exam Questions and Answers to help you pass Microsoft and other Hot exam 100%!

2014 Latest Oracle 1Z0-060 Exam Demo Free Download!

QUESTION 1
You notice that the performance of your production 24/7 Oracle database significantly degraded. Sometimes you are not able to connect to the instance because it hangs. You do not want to restart the database instance.
How can you detect the cause of the degraded performance?

A.    Enable Memory Access Mode, which reads performance data from SGA.
B.    Use emergency monitoring to fetch data directly from SGA analysis.
C.    Run Automatic Database Diagnostic Monitor (ADDM) to fetch information from the latest Automatic
Workload Repository (AWR) snapshots.
D.    Use Active Session History (ASH) data and hang analysis in regular performance monitoring.
E.    Run ADDM in diagnostic mode.

Answer: C
Explanation:
*In most cases, ADDM output should be the first place that a DBA looks when notified of a performance problem.
*Performance degradation of the database occurs when your database was performing optimally in the past, such as 6 months ago, but has gradually degraded to a point where it becomes noticeable to the users. The Automatic Workload Repository (AWR) Compare Periods report enables you to compare database performance between two periods of time.
While an AWR report shows AWR data between two snapshots (or two points in time), the AWR Compare Periods report shows the difference between two periods (or two AWR reports with a total of four snapshots). Using the AWR Compare Periods report helps you to identify detailed performance attributes and configuration settings that differ between two time periods.
Reference:Resolving Performance Degradation Over Time

QUESTION 2
You plan to use the In Database Archiving feature of Oracle Database 12c, and store rows that are inactive for over three months, in Hybrid Columnar Compressed (HCC) format.
Which three storage options support the use of HCC?

A.    ASM disk groups with ASM disks consisting of Exadata Grid Disks.
B.    ASM disk groups with ASM disks consisting of LUNS on any Storage Area Network array
C.    ASM disk groups with ASM disks consisting of any zero padded NFS-mounted files
D.    Database files stored in ZFS and accessed using conventional NFS mounts.
E.    Database files stored in ZFS and accessed using the Oracle Direct NFS feature
F.    Database files stored in any file system and accessed using the Oracle Direct NFS feature
G.    ASM disk groups with ASM disks consisting of LUNs on Pillar Axiom Storage arrays

Answer: AEG
Explanation:
HCC requires the use of Oracle Storage ?Exadata(A), Pillar Axiom(G)or Sun ZFS Storage Appliance (ZFSSA).
Note:
*Hybrid Columnar Compression, initially only available on Exadata, has been extended to supportPillar Axiom and Sun ZFS Storage Appliance (ZFSSA) storage when used with Oracle Database Enterprise Edition 11.2.0.3 and above
*Oracle offers the ability to manage NFS using a feature called Oracle Direct NFS (dNFS). Oracle Direct NFS implements NFS V3 protocol within the Oracle database kernel itself. Oracle Direct NFS client overcomes many of the challenges associated with using NFS with the Oracle Database with simple configuration, betterperformance than traditional NFS clients, and offers consistent configuration across platforms.
QUESTION 3
In your multitenant container database (CDB) containing pluggable databases (PDB), users complain about performance degradation.
How does real-time Automatic database Diagnostic Monitor (ADDM) check performance degradation and provide solutions?

A.    It collects data from SGA and compares it with a preserved snapshot.
B.    It collects data from SGA, analyzes it, and provides a report.
C.    It collects data from SGA and compares it with the latest snapshot.
D.    It collects data from both SGA and PGA, analyzes it, and provides a report.

Answer: B
Explanation:
Note:
*The multitenant architecture enables an Oracle database to function as a multitenant container database (CDB) that includes zero, one, or many customer-created pluggable databases (PDBs). A PDB is a portable collection of schemas, schema objects, and nonschema objects that appears to an Oracle Net client as a non-CDB. All Oracle databases before Oracle Database 12c were non-CDBs.
*The System Global Area (SGA) is a group of shared memory areas that are dedicated to an Oracle "instance" (an instance is your database programs and RAM). *The PGA (Program or Process Global Area) is a memory area (RAM) that stores data and control information for a single process.

QUESTION 4
The tnsnames.ora file has an entry for the service alias ORCL as follows:
  clip_image002[4]
The TNS ping command executes successfully when tested with ORCL; however, from the same OS user session, you are not able to connect to the database instance with the following command:
SQL > CONNECT scott/tiger@orcl
What could be the reason for this?

A.    The listener is not running on the database node.
B.    The TNS_ADMIN environment variable is set to the wrong value.
C.    The orcl.oracle.com database service is not registered with the listener.
D.    The DEFAULT_DOMAIN parameter is set to the wrong value in the sqlnet.ora file.
E.    The listener is running on a different port.

Answer: C
Explanation:
Service registration enables the listener to determine whether a database service and its service handlers are available. A service handler is a dedicated server process or dispatcher that acts as a connection point to a database. During registration, the LREG process provides the listener with the instance name, database service names, and the type and addresses of service handlers. This information enables the listener to start a service handler when a client request arrives.

QUESTION 5
Examine the following steps of privilege analysis for checking and revoking excessive, unused privileges granted to users:
1. Create a policy to capture the privilege used by a user for privilege analysis.
2. Generate a report with the data captured for a specified privilege capture.
3. Start analyzing the data captured by the policy.
4. Revoke the unused privileges.
5. Compare the used and unused privileges’ lists.
6. Stop analyzing the data.
Identify the correct sequence of steps.

A.    1, 3, 5, 6, 2, 4
B.    1, 3, 6, 2, 5, 4
C.    1, 3, 2, 5, 6, 4
D.    1, 3, 2, 5, 6, 4
E.    1, 3, 5, 2, 6, 4

Answer: B
Explanation:
1.Create a policy to capture the privilege used by a user for privilege analysis.
3.Start analyzing the data captured by the policy.
6.Stop analyzing the data.
2.Generate a report with the data captured for a specified privilege capture.
5.Compare the used and unused privileges’ lists.
4.Revoke the unused privileges.

QUESTION 6
You database is running an ARCHIVELOG mode.
The following parameter are set in your database instance:
LOG_ARCHIVE_FORMAT = arch+%t_%r.arc
LOG_ARCHIVE_DEST_1 = `LOCATION = /disk1/archive’
DB_RECOVERY_FILE_DEST_SIZE = 50G
DB_RECOVERY_FILE = `/u01/oradata’
Which statement is true about the archived redo log files?

A.    They are created only in the location specified by the LOG_ARCHIVE_DEST_1 parameter.
B.    They are created only in the Fast Recovery Area.
C.    They are created in the location specified by the LOG_ARCHIVE_DEST_1 parameter and in the
default location $ORACLE_HOME/dbs/arch.
D.    They are created in the location specified by the LOG_ARCHIVE_DEST_1 parameter and the location
specified by the DB_RECOVERY_FILE_DEST parameter.

Answer: A
Explanation:
You can choose to archive redo logs to a single destination or to multiple destinations.
Destinations can be local–within the local file system or an Oracle Automatic Storage Management (Oracle ASM) disk group–or remote (on a standby database). When you archive to multiple destinations, a copy of each filled redo log file is written to each destination. These redundant copies help ensure that archived logs are always available in the event of a failure at one of the destinations. To archive to only a single destination, specify that destination using the LOG_ARCHIVE_DESTandLOG_ARCHIVE_DUPLEX_DESTinitialization parameters. ARCHIVE_DEST initialization parameter. To archive to multiple destinations, you can choose to archive to two or more locations using the LOG_ARCHIVE_DEST_n initialization parameters, or to archive only to a primary and secondary destination using the LOG_ ARCHIVE_DEST and LOG_ARCHIVE_DUPLEX_DEST initialization parameters.

QUESTION 7
Your multitenant container database (CDB) is running in ARCHIVELOG mode. You connect to the CDB RMAN. Examine the following command and its output:
 clip_image002[6]
You execute the following command:
RMAN > BACKUP DATABASE PLUS ARCHIVELOG;
Which data files will be backed up?

A.    Data files that belong to only the root container
B.    Data files that belong to the root container and all the pluggable databases (PDBs)
C.    Data files that belong to only the root container and PDB$SEED
D.    Data files that belong to the root container and all the PDBs excluding PDB$SEED

Answer: B
Explanation:
Backing Up a Whole CDB
Backing up a whole CDB is similar to backing up a non-CDB. When you back up a whole CDB, RMAN backs up the root, all the PDBs, and the archived redo logs. You can then recover either the whole CDB, the root only, or one or more PDBs from the CDB backup.
Note:
*You can back up and recover a whole CDB, the root only, or one or more PDBs.
*Backing Up Archived Redo Logs with RMAN
Archived redo logs are the key to successful media recovery. Back them up regularly. You can back up logs with BACKUP ARCHIVELOG, or back up logs while backing up datafiles and control files by specifying BACKUP … PLUS ARCHIVELOG.

QUESTION 8
You are administering a database stored in Automatic Storage management (ASM). The files are stored in the DATA disk group. You execute the following command:
SQL > ALTER DISKGROUP data ADD ALIAS `+data/prod/myfile.dbf’ FOR `+data.231.45678′;
What is the result?
A.    The file `+data.231.54769′ is physically relocated to `+data/prod’ and renamed as `myfile.dbf’.
B.    The file `+data.231.54769′ is renamed as `myfile.dbf’, and copied to `+data/prod’.
C.    The file `+data.231.54769′ remains in the same location andasynonym’myfile.dbf’ is created.
D.    The file `myfile.dbf’ is created in `+data/prod’ and the reference to `+data.231.54769′ in the data
dictionary removed.

Answer: C
Explanation:
ADD ALIAS
Use this clause to create an alias name for an Oracle ASM filename. The alias_name consists of the full directory path and the alias itself.

QUESTION 9
Which three functions are performed by the SQL Tuning Advisor?

A.    Building and implementing SQL profiles
B.    Recommending the optimization of materialized views
C.    Checking query objects for missing and stale statistics
D.    Recommending bitmap, function-based, and B-tree indexes
E.    Recommending the restructuring of SQL queries that are using bad plans

Answer: ACE
Explanation:
The SQL Tuning Advisor takes one or more SQL statements as an input and invokes the Automatic Tuning Optimizer to perform SQL tuning on the statements. The output of the SQL Tuning Advisor is in the form of an advice or recommendations, along with a rationale for eachrecommendation and its expected benefit. The recommendation relates to collection of statistics on objects(C), creation of new indexes, restructuring of the SQL statement(E), or creation of a SQL profile(A). You can choose to accept the recommendation to complete the tuning of the SQL statements.

QUESTION 10
Examine the following command:
ALTER SYSTEM SET enable_ddl_logging=FALSE;
Which statement is true?

A.    None of the data definition language (DDL) statements are logged in the trace file.
B.    Only DDL commands that resulted in errors are logged in the alert log file.
C.    A new log.xml file that contains the DDL statements is created, and the DDL command details are
removed from the alert log file.
D.    Only DDL commands that resulted in the creation of new database files are logged.

Answer: A
Explanation:
ENABLE_DDL_LOGGING enables or disables the writing of a subset of data definition language (DDL) statements to a DDL alert log.
The DDL log is a file that has the same format and basic behavior as the alert log, but it only contains the DDL statements issued by the database. The DDL log is created only for the RDBMS component and only if the ENABLE_DDL_LOGGING initialization parameter is set to true. When this parameter is set to false, DDL statements are not included in any log.

QUESTION 11
Your multitenant container database (CDB) contains three pluggable database (PDBs). You find that the control file is damaged. You plan to use RMAN to recover the control file. There are no startup triggers associated with the PDBs.
Which three steps should you perform to recover the control file and make the database fully operational?

A.    Mount the container database (CDB) and restore the control file from the control file auto backup.
B.    Recover and open the CDB in NORMAL mode.
C.    Mount the CDB and then recover and open the database, with the RESETLOGS option.
D.    Open all the pluggable databases.
E.    Recover each pluggable database.
F.    Start the database instance in the nomount stage and restore the control file from control file auto backup.

Answer: CDF
Explanation:
Step 1: F
Step 2: D
Step 3: C: If all copies of the current control file are lost or damaged, then you must restore and mount a backup control file. You must then run the RECOVERcommand, even if no data files have been restored, and open the database with the RESETLOGS option.
Note:
*RMAN and Oracle Enterprise Manager Cloud Control (Cloud Control) provide full support for backup and recovery in a multitenant environment. You can back up and recover a whole multitenant container database (CDB), root only, or one or more pluggable databases (PDBs).

Passing your Oracle 1Z0-060 Exam by using the latest Oracle 1Z0-060 Exam Demo Full Version: http://www.braindump2go.com/1z0-060.html

, , , ,

Comments are currently closed.