Winter Special Sale - Limited Time 60% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 575363r9

Welcome To DumpsPedia

CCA-500 Sample Questions Answers

Questions 4

You observed that the number of spilled records from Map tasks far exceeds the number of map output records. Your child heap size is 1GB and your io.sort.mb value is set to 1000MB. How would you tune your io.sort.mb value to achieve maximum memory to disk I/O ratio?

Options:

A.

For a 1GB child heap size an io.sort.mb of 128 MB will always maximize memory to disk I/O

B.

Increase the io.sort.mb to 1GB

C.

Decrease the io.sort.mb value to 0

D.

Tune the io.sort.mb value until you observe that the number of spilled records equals (or is as close to equals) the number of map output records.

Buy Now
Questions 5

Your cluster is running MapReduce version 2 (MRv2) on YARN. Your ResourceManager is configured to use the FairScheduler. Now you want to configure your scheduler such that a new user on the cluster can submit jobs into their own queue application submission. Which configuration should you set?

Options:

A.

You can specify new queue name when user submits a job and new queue can be created dynamically if the property yarn.scheduler.fair.allow-undecleared-pools = true

B.

Yarn.scheduler.fair.user.fair-as-default-queue = false and yarn.scheduler.fair.allow-undecleared-pools = true

C.

You can specify new queue name when user submits a job and new queue can be created dynamically if yarn .schedule.fair.user-as-default-queue = false

D.

You can specify new queue name per application in allocations.xml file and have new jobs automatically assigned to the application queue

Buy Now
Questions 6

In CDH4 and later, which file contains a serialized form of all the directory and files inodes in the filesystem, giving the NameNode a persistent checkpoint of the filesystem metadata?

Options:

A.

fstime

B.

VERSION

C.

Fsimage_N (where N reflects transactions up to transaction ID N)

D.

Edits_N-M (where N-M transactions between transaction ID N and transaction ID N)

Buy Now
Questions 7

A slave node in your cluster has 4 TB hard drives installed (4 x 2TB). The DataNode is configured to store HDFS blocks on all disks. You set the value of the dfs.datanode.du.reserved parameter to 100 GB. How does this alter HDFS block storage?

Options:

A.

25GB on each hard drive may not be used to store HDFS blocks

B.

100GB on each hard drive may not be used to store HDFS blocks

C.

All hard drives may be used to store HDFS blocks as long as at least 100 GB in total is available on the node

D.

A maximum if 100 GB on each hard drive may be used to store HDFS blocks

Buy Now
Questions 8

Which two are features of Hadoop’s rack topology? (Choose two)

Options:

A.

Configuration of rack awareness is accomplished using a configuration file. You cannot use a rack topology script.

B.

Hadoop gives preference to intra-rack data transfer in order to conserve bandwidth

C.

Rack location is considered in the HDFS block placement policy

D.

HDFS is rack aware but MapReduce daemon are not

E.

Even for small clusters on a single rack, configuring rack awareness will improve performance

Buy Now
Questions 9

Your company stores user profile records in an OLTP databases. You want to join these records with web server logs you have already ingested into the Hadoop file system. What is the best way to obtain and ingest these user records?

Options:

A.

Ingest with Hadoop streaming

B.

Ingest using Hive’s IQAD DATA command

C.

Ingest with sqoop import

D.

Ingest with Pig’s LOAD command

E.

Ingest using the HDFS put command

Buy Now
Questions 10

What two processes must you do if you are running a Hadoop cluster with a single NameNode and six DataNodes, and you want to change a configuration parameter so that it affects all six DataNodes. (Choose two)

Options:

A.

You must modify the configuration files on the NameNode only. DataNodes read their configuration from the master nodes

B.

You must modify the configuration files on each of the DataNodes machines

C.

You don’t need to restart any daemon, as they will pick up changes automatically

D.

You must restart the NameNode daemon to apply the changes to the cluster

E.

You must restart all six DatNode daemon to apply the changes to the cluster

Buy Now
Questions 11

On a cluster running MapReduce v2 (MRv2) on YARN, a MapReduce job is given a directory of 10 plain text files as its input directory. Each file is made up of 3 HDFS blocks. How many Mappers will run?

Options:

A.

We cannot say; the number of Mappers is determined by the ResourceManager

B.

We cannot say; the number of Mappers is determined by the developer

C.

30

D.

3

E.

10

F.

We cannot say; the number of mappers is determined by the ApplicationMaster

Buy Now
Questions 12

Table schemas in Hive are:

Options:

A.

Stored as metadata on the NameNode

B.

Stored along with the data in HDFS

C.

Stored in the Metadata

D.

Stored in ZooKeeper

Questions 13

You are running a Hadoop cluster with a NameNode on host mynamenode. What are two ways to determine available HDFS space in your cluster?

Options:

A.

Run hdfs fs –du / and locate the DFS Remaining value

B.

Run hdfs dfsadmin –report and locate the DFS Remaining value

C.

Run hdfs dfs / and subtract NDFS Used from configured Capacity

D.

Connect to http://mynamenode:50070/dfshealth.jsp and locate the DFS remaining value

Buy Now
Questions 14

Which YARN daemon or service negotiations map and reduce Containers from the Scheduler, tracking their status and monitoring progress?

Options:

A.

NodeManager

B.

ApplicationMaster

C.

ApplicationManager

D.

ResourceManager

Buy Now
Questions 15

You are migrating a cluster from MApReduce version 1 (MRv1) to MapReduce version 2 (MRv2) on YARN. You want to maintain your MRv1 TaskTracker slot capacities when you migrate. What should you do/

Options:

A.

Configure yarn.applicationmaster.resource.memory-mb and yarn.applicationmaster.resource.cpu-vcores so that ApplicationMaster container allocations match the capacity you require.

B.

You don’t need to configure or balance these properties in YARN as YARN dynamically balances resource management capabilities on your cluster

C.

Configure mapred.tasktracker.map.tasks.maximum and mapred.tasktracker.reduce.tasks.maximum ub yarn-site.xml to match your cluster’s capacity set by the yarn-scheduler.minimum-allocation

D.

Configure yarn.nodemanager.resource.memory-mb and yarn.nodemanager.resource.cpu-vcores to match the capacity you require under YARN for each NodeManager

Buy Now
Questions 16

You’re upgrading a Hadoop cluster from HDFS and MapReduce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce a block size of 128MB for all new files written to the cluster after upgrade. What should you do?

Options:

A.

You cannot enforce this, since client code can always override this value

B.

Set dfs.block.size to 128M on all the worker nodes, on all client machines, and on the NameNode, and set the parameter to final

C.

Set dfs.block.size to 128 M on all the worker nodes and client machines, and set the parameter to final. You do not need to set this value on the NameNode

D.

Set dfs.block.size to 134217728 on all the worker nodes, on all client machines, and on the NameNode, and set the parameter to final

E.

Set dfs.block.size to 134217728 on all the worker nodes and client machines, and set the parameter to final. You do not need to set this value on the NameNode

Buy Now
Questions 17

You are planning a Hadoop cluster and considering implementing 10 Gigabit Ethernet as the network fabric. Which workloads benefit the most from faster network fabric?

Options:

A.

When your workload generates a large amount of output data, significantly larger than the amount of intermediate data

B.

When your workload consumes a large amount of input data, relative to the entire capacity if HDFS

C.

When your workload consists of processor-intensive tasks

D.

When your workload generates a large amount of intermediate data, on the order of the input data itself

Buy Now
Questions 18

You have recently converted your Hadoop cluster from a MapReduce 1 (MRv1) architecture to MapReduce 2 (MRv2) on YARN architecture. Your developers are accustomed to specifying map and reduce tasks (resource allocation) tasks when they run jobs: A developer wants to know how specify to reduce tasks when a specific job runs. Which method should you tell that developers to implement?

Options:

A.

MapReduce version 2 (MRv2) on YARN abstracts resource allocation away from the idea of “tasks” into memory and virtual cores, thus eliminating the need for a developer to specify the number of reduce tasks, and indeed preventing the developer from specifying the number of reduce tasks.

B.

In YARN, resource allocations is a function of megabytes of memory in multiples of 1024mb. Thus, they should specify the amount of memory resource they need by executing –D mapreduce-reduces.memory-mb-2048

C.

In YARN, the ApplicationMaster is responsible for requesting the resource required for a specific launch. Thus, executing –D yarn.applicationmaster.reduce.tasks=2 will specify that the ApplicationMaster launch two task contains on the worker nodes.

D.

Developers specify reduce tasks in the exact same way for both MapReduce version 1 (MRv1) and MapReduce version 2 (MRv2) on YARN. Thus, executing –D mapreduce.job.reduces-2 will specify reduce tasks.

E.

In YARN, resource allocation is function of virtual cores specified by the ApplicationManager making requests to the NodeManager where a reduce task is handeled by a single container (and thus a single virtual core). Thus, the developer needs to specify the number of virtual cores to the NodeManager by executing –p yarn.nodemanager.cpu-vcores=2

Buy Now
Exam Code: CCA-500
Exam Name: Cloudera Certified Administrator for Apache Hadoop (CCAH)
Last Update: Nov 20, 2024
Questions: 60
$64  $159.99
$48  $119.99
$40  $99.99
buy now CCA-500