Network Security Internet Technology Development Database Servers Mobile Phone Android Software Apple Software Computer Software News IT Information

In addition to Weibo, there is also WeChat

Please pay attention

WeChat public account

Shulou

What are the Hadoop interview questions?

2025-03-26 Update From: SLTechnology News&Howtos shulou NAV: SLTechnology News&Howtos > Internet Technology >

Share

Shulou(Shulou.com)06/01 Report--

This article mainly explains "what are the Hadoop interview questions". The explanation in the article is simple and clear, and it is easy to learn and understand. Please follow the editor's train of thought to study and learn "what are the Hadoop interview questions"?

Single multiple choice question

1. Which of the following programs is responsible for HDFS data storage?

A) NameNode

B) Jobtracker

C) Datanode

D) secondaryNameNode

E) tasktracker

2. How many copies of block in HDfS are saved by default?

A) 3 copies

B) 2 copies

C) 1 copy

D) uncertainty

3. Which of the following programs is usually started on the same node as NameNode?

A) SecondaryNameNode

B) DataNode

C) TaskTracker

D) Jobtracker

4. Hadoop author

A) Martin Fowler

B) Kent Beck

C) Doug cutting

5. HDFS default Block Size

A) 32MB

B) 64MB

C) 128MB

6. Which of the following is usually the main bottleneck of a cluster?

A) CPU

B) Network

C) disk

D) memory

7. Which is true about SecondaryNameNode?

A) it is a hot backup for NameNode

B) it has no memory requirements

C) its purpose is to help NameNode merge and edit logs and reduce NameNode startup time

D) SecondaryNameNode should be deployed to one node with NameNode

Multiple choice topics:

8. Which of the following can be used as a cluster management tool

A) Puppet

B) Pdsh

C) Cloudera Manager

D) d) Zookeeper

9. Which of the following is correct to configure rack awareness?

A) if there is a problem with a rack, it will not affect data reading and writing

B) when writing data, it will be written to the DataNode of different racks.

C) MapReduce will get the network data close to it according to the rack

10. Which of the following is correct when uploading files on Client?

A) data is passed to DataNode via NameNode

B) the Client side cuts the file into Block, and uploads the file in turn

C) Client only uploads data to one DataNode, and then NameNode is responsible for Block replication

11. Which of the following is the mode in which Hadoop runs

A) standalone version

B) pseudo-distributed

C) distributed

12. What are the ways for Cloudera to install CDH

A) Cloudera manager

B) Tar ball

C) Yum d) Rpm

Judgment question:

13. Ganglia can not only monitor, but also alarm. ()

14. Block Size cannot be modified. ()

15. Nagios cannot monitor Hadoop clusters because it does not provide Hadoop support. ()

16. If NameNode terminates unexpectedly, SecondaryNameNode will replace it to keep the cluster working. ()

17. Cloudera CDH needs to be used for a fee. ()

18. Hadoop is developed by Java, so MapReduce only supports writing in Java. ()

19. Hadoop supports random reading and writing of data. ()

20. NameNode is responsible for managing every read and write request on the metadata,client side. It will read or write metadata information from the disk and feed back to the client side. ()

21. The NameNode local disk holds the location information of the Block. ()

twenty-two。 DataNode maintains communication with NameNode over a persistent connection. ()

23. Hadoop itself has strict rights management and security measures to ensure the normal operation of the cluster. ()

24. The Slave node stores data, so the larger its disk, the better. ()

25. The hadoop dfsadmin-report command is used to detect HDFS damaged blocks. ()

twenty-six。 The default scheduler policy for Hadoop is FIFO ()

twenty-seven。 Each node in the cluster should be equipped with RAID to avoid damage to a single disk and affect the operation of the entire node. ()

twenty-eight。 Because there are multiple copies of HDFS, there is no single point of problem with NameNode. ()

twenty-nine。 Each map slot is a thread. ()

thirty。 Mapreduce's input split is a block. ()

thirty-one。 The Web UI port of NameNode is 50030, and it starts the Web service through jetty. ()

thirty-two。 The HADOOP_HEAPSIZE in the Hadoop environment variable is used to set memory for all Hadoop daemon threads. It defaults to 200 GB. ()

thirty-three。 When DataNode joins cluster, if an incompatible file version is reported in log, NameNode is required to perform a "Hadoop namenode-format" operation to format the disk. ()

Thank you for your reading, the above is the content of "what are the Hadoop interview questions?" after the study of this article, I believe you have a deeper understanding of what the Hadoop interview questions have, and the specific use needs to be verified in practice. Here is, the editor will push for you more related knowledge points of the article, welcome to follow!

Welcome to subscribe "Shulou Technology Information " to get latest news, interesting things and hot topics in the IT industry, and controls the hottest and latest Internet news, technology news and IT industry trends.

Views: 0

*The comments in the above article only represent the author's personal views and do not represent the views and positions of this website. If you have more insights, please feel free to contribute and share.

Share To

Internet Technology

Wechat

© 2024 shulou.com SLNews company. All rights reserved.

12
Report