Hadoop Big Data Online Test

Table of Contents

1. Which of the following is a key component of the Hadoop ecosystem?

A. HDFS
B. Oracle
C. MongoDB
D. PostgreSQL

Answer: A

2. What is the main purpose of Hadoop Distributed File System (HDFS)?

A. To store relational data in tables
B. To process real-time data streams
C. To store large datasets across multiple machines
D. To manage SQL-based queries

Answer: C

3. Which component of Hadoop is responsible for processing data?

A. HDFS
B. MapReduce
C. HBase
D. Yarn

Answer: B

4. What is the function of NameNode in HDFS?

A. Stores actual data blocks
B. Manages metadata and file system namespace
C. Executes Map tasks
D. Manages data replication

Answer: B

5. What is the role of DataNode in Hadoop?

A. Stores metadata
B. Stores actual data blocks
C. Monitors MapReduce jobs
D. Executes SQL queries

Answer: B

6. What is the default block size in Hadoop 3.x?

A. 64 MB
B. 128 MB
C. 256 MB
D. 512 MB

Answer: B

7. Which of the following is a resource management layer in Hadoop?

A. YARN
B. Pig
C. Sqoop
D. Oozie

Answer: A

8. Which language is primarily used for writing Hadoop MapReduce programs?

A. Python
B. Java
C. SQL
D. C++

Answer: B

9. What does MapReduce consist of?

A. Mapper and Combiner
B. Mapper and Reducer
C. Mapper only
D. Reducer only

Answer: B

10. What is a “combiner” in Hadoop MapReduce?

A. A backup reducer
B. A pre-reducer that performs local aggregation
C. A secondary mapper
D. A task scheduler

Answer: B

11. What does YARN stand for?

A. Yet Another Recursive NameNode
B. Yet Another Resource Negotiator
C. Your Advanced Resource Network
D. Yearly Assigned Resource Node

Answer: B

12. Which tool is used for data ingestion from RDBMS to Hadoop?

A. Pig
B. Hive
C. Sqoop
D. Flume

Answer: C

13. Which Hadoop ecosystem tool is used for real-time data ingestion?

A. Hive
B. Oozie
C. Flume
D. Sqoop

Answer: C

14. What type of language is Pig Latin in Hadoop?

A. Declarative
B. Procedural
C. Object-oriented
D. Functional

Answer: B

15. What is Hive mainly used for?

A. Workflow scheduling
B. Streaming analytics
C. Data warehousing and SQL-like queries
D. Data ingestion

Answer: C

16. What file format is best for columnar storage in Hadoop?

A. CSV
B. JSON
C. ORC
D. TXT

Answer: C

17. Which command is used to copy data from local to HDFS?

A. hadoop fs -get
B. hadoop fs -put
C. hadoop fs -delete
D. hadoop fs -ls

Answer: B

18. What happens if a DataNode fails in Hadoop?

A. The system stops
B. NameNode replaces it automatically
C. Data is replicated from another DataNode
D. All data is lost

Answer: C

19. Which of the following is a scheduling component in Hadoop?

A. ResourceManager
B. NodeManager
C. DataNode
D. NameNode

Answer: A

20. What is the major advantage of Hadoop?

A. High licensing cost
B. Centralized storage
C. Scalability and fault tolerance
D. Requires single-node architecture

Answer: C

Share this article

Enroll Free demo class
Enroll IT Courses

Enroll Free demo class

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Join Free Demo Class

Let's have a chat