Black Friday Special - 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: dm70dm

Hot Vendors
Note! The HDPCD Exam is no longer available.

HDPCD Hortonworks Data Platform Certified Developer Questions and Answers

Questions 4

Which Hadoop component is responsible for managing the distributed file system metadata?

Options:

A.

NameNode

B.

Metanode

C.

DataNode

D.

NameSpaceManager

Buy Now
Questions 5

In a MapReduce job, the reducer receives all values associated with same key. Which statement best describes the ordering of these values?

Options:

A.

The values are in sorted order.

B.

The values are arbitrarily ordered, and the ordering may vary from run to run of the same MapReduce job.

C.

The values are arbitrary ordered, but multiple runs of the same MapReduce job will always have the same ordering.

D.

Since the values come from mapper outputs, the reducers will receive contiguous sections of sorted values.

Buy Now
Questions 6

You want to perform analysis on a large collection of images. You want to store this data in HDFS and process it with MapReduce but you also want to give your data analysts and data scientists the ability to process the data directly from HDFS with an interpreted high-level programming language like Python. Which format should you use to store this data in HDFS?

Options:

A.

SequenceFiles

B.

Avro

C.

JSON

D.

HTML

E.

XML

F.

CSV

Buy Now
Questions 7

Given the following Pig command:

logevents = LOAD 'input/my.log' AS (date:chararray, levehstring, code:int, message:string);

Which one of the following statements is true?

Options:

A.

The logevents relation represents the data from the my.log file, using a comma as the parsing delimiter

B.

The logevents relation represents the data from the my.log file, using a tab as the parsing delimiter

C.

The first field of logevents must be a properly-formatted date string or table return an error

D.

The statement is not a valid Pig command

Buy Now
Questions 8

When is the earliest point at which the reduce method of a given Reducer can be called?

Options:

A.

As soon as at least one mapper has finished processing its input split.

B.

As soon as a mapper has emitted at least one record.

C.

Not until all mappers have finished processing all records.

D.

It depends on the InputFormat used for the job.

Buy Now
Questions 9

You need to run the same job many times with minor variations. Rather than hardcoding all job configuration options in your drive code, you’ve decided to have your Driver subclass org.apache.hadoop.conf.Configured and implement the org.apache.hadoop.util.Tool interface.

Indentify which invocation correctly passes.mapred.job.name with a value of Example to Hadoop?

Options:

A.

hadoop “mapred.job.name=Example” MyDriver input output

B.

hadoop MyDriver mapred.job.name=Example input output

C.

hadoop MyDrive –D mapred.job.name=Example input output

D.

hadoop setproperty mapred.job.name=Example MyDriver input output

E.

hadoop setproperty (“mapred.job.name=Example”) MyDriver input output

Buy Now
Questions 10

For each intermediate key, each reducer task can emit:

Options:

A.

As many final key-value pairs as desired. There are no restrictions on the types of those key-value pairs (i.e., they can be heterogeneous).

B.

As many final key-value pairs as desired, but they must have the same type as the intermediate key-value pairs.

C.

As many final key-value pairs as desired, as long as all the keys have the same type and all the values have the same type.

D.

One final key-value pair per value associated with the key; no restrictions on the type.

E.

One final key-value pair per key; no restrictions on the type.

Buy Now
Questions 11

What does the following WebHDFS command do?

Curl -1 -L <a href="“https://host:port/webhdfs/v1/foo/bar?op=OPEN”">“https://host:port/webhdfs/v1/foo/bar?op=OPEN”

Options:

A.

Make a directory /foo/bar

B.

Read a file /foo/bar

C.

List a directory /foo

D.

Delete a directory /foo/bar

Buy Now
Questions 12

You need to perform statistical analysis in your MapReduce job and would like to call methods in the Apache Commons Math library, which is distributed as a 1.3 megabyte Java archive (JAR) file. Which is the best way to make this library available to your MapReducer job at runtime?

Options:

A.

Have your system administrator copy the JAR to all nodes in the cluster and set its location in the HADOOP_CLASSPATH environment variable before you submit your job.

B.

Have your system administrator place the JAR file on a Web server accessible to all cluster nodes and then set the HTTP_JAR_URL environment variable to its location.

C.

When submitting the job on the command line, specify the –libjars option followed by the JAR file path.

D.

Package your code and the Apache Commands Math library into a zip file named JobJar.zip

Buy Now
Questions 13

You have written a Mapper which invokes the following five calls to the OutputColletor.collect method:

output.collect (new Text (“Apple”), new Text (“Red”) ) ;

output.collect (new Text (“Banana”), new Text (“Yellow”) ) ;

output.collect (new Text (“Apple”), new Text (“Yellow”) ) ;

output.collect (new Text (“Cherry”), new Text (“Red”) ) ;

output.collect (new Text (“Apple”), new Text (“Green”) ) ;

How many times will the Reducer’s reduce method be invoked?

Options:

A.

6

B.

3

C.

1

D.

0

E.

5

Buy Now
Questions 14

Which one of the following statements describes a Hive user-defined aggregate function?

Options:

A.

Operates on multiple input rows and creates a single row as output

B.

Operates on a single input row and produces a single row as output

C.

Operates on a single input row and produces a table as output

D.

Operates on multiple input rows and produces a table as output

Buy Now
Questions 15

You have just executed a MapReduce job. Where is intermediate data written to after being emitted from the Mapper’s map method?

Options:

A.

Intermediate data in streamed across the network from Mapper to the Reduce and is never written to disk.

B.

Into in-memory buffers on the TaskTracker node running the Mapper that spill over and are written into HDFS.

C.

Into in-memory buffers that spill over to the local file system of the TaskTracker node running the Mapper.

D.

Into in-memory buffers that spill over to the local file system (outside HDFS) of the TaskTracker node running the Reducer

E.

Into in-memory buffers on the TaskTracker node running the Reducer that spill over and are written into HDFS.

Buy Now
Questions 16

Identify the tool best suited to import a portion of a relational database every day as files into HDFS, and generate Java classes to interact with that imported data?

Options:

A.

Oozie

B.

Flume

C.

Pig

D.

Hue

E.

Hive

F.

Sqoop

G.

fuse-dfs

Buy Now
Exam Code: HDPCD
Exam Name: Hortonworks Data Platform Certified Developer
Last Update: Mar 21, 2024
Questions: 108
dumpsmate guaranteed to pass
24/7 Customer Support

DumpsMate's team of experts is always available to respond your queries on exam preparation. Get professional answers on any topic of the certification syllabus. Our experts will thoroughly satisfy you.

Site Secure

mcafee secure

TESTED 23 Nov 2024