Winter Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: bigdisc65

CCA-500 Exam Dumps - Cloudera CCAH Questions and Answers

Page: 1 / 2
Questions 4

What two processes must you do if you are running a Hadoop cluster with a single NameNode and six DataNodes, and you want to change a configuration parameter so that it affects all six DataNodes. (Choose two)

Options:

A.

You must modify the configuration files on the NameNode only. DataNodes read their configuration from the master nodes

B.

You must modify the configuration files on each of the DataNodes machines

C.

You don’t need to restart any daemon, as they will pick up changes automatically

D.

You must restart the NameNode daemon to apply the changes to the cluster

E.

You must restart all six DatNode daemon to apply the changes to the cluster

Buy Now
Questions 5

You’re upgrading a Hadoop cluster from HDFS and MapReduce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce a block size of 128MB for all new files written to the cluster after upgrade. What should you do?

Options:

A.

You cannot enforce this, since client code can always override this value

B.

Set dfs.block.size to 128M on all the worker nodes, on all client machines, and on the NameNode, and set the parameter to final

C.

Set dfs.block.size to 128 M on all the worker nodes and client machines, and set the parameter to final. You do not need to set this value on the NameNode

D.

Set dfs.block.size to 134217728 on all the worker nodes, on all client machines, and on the NameNode, and set the parameter to final

E.

Set dfs.block.size to 134217728 on all the worker nodes and client machines, and set the parameter to final. You do not need to set this value on the NameNode

Buy Now
Questions 6

Your cluster’s mapred-start.xml includes the following parameters

mapreduce.map.memory.mb

4096

mapreduce.reduce.memory.mb

8192

And any cluster’s yarn-site.xml includes the following parameters

yarn.nodemanager.vmen-pmen-ration

2.1

What is the maximum amount of virtual memory allocated for each map task before YARN will kill its Container?

Options:

A.

4 GB

B.

17.2 GB

C.

8.9 GB

D.

8.2 GB

E.

24.6 GB

Buy Now
Questions 7

You want to understand more about how users browse your public website. For example, you want to know which pages they visit prior to placing an order. You have a server farm of 200 web servers hosting your website. Which is the most efficient process to gather these web server across logs into your Hadoop cluster analysis?

Options:

A.

Sample the web server logs web servers and copy them into HDFS using curl

B.

Ingest the server web logs into HDFS using Flume

C.

Channel these clickstreams into Hadoop using Hadoop Streaming

D.

Import all user clicks from your OLTP databases into Hadoop using Sqoop

E.

Write a MapReeeduce job with the web servers for mappers and the Hadoop cluster nodes for reducers

Buy Now
Page: 1 / 2
Exam Code: CCA-500
Exam Name: Cloudera Certified Administrator for Apache Hadoop (CCAH)
Last Update: Nov 21, 2024
Questions: 60
CCA-500 pdf

CCA-500 PDF

$28  $80
CCA-500 Engine

CCA-500 Testing Engine

$33.25  $95
CCA-500 PDF + Engine

CCA-500 PDF + Testing Engine

$45.5  $130