1. Which of the following is a key component of the Hadoop ecosystem?
A. HDFS
B. Oracle
C. MongoDB
D. PostgreSQL
Answer: A
2. What is the main purpose of Hadoop Distributed File System (HDFS)?
A. To store relational data in tables
B. To process real-time data streams
C. To store large datasets across multiple machines
D. To manage SQL-based queries
Answer: C
3. Which component of Hadoop is responsible for processing data?
A. HDFS
B. MapReduce
C. HBase
D. Yarn
Answer: B
4. What is the function of NameNode in HDFS?
A. Stores actual data blocks
B. Manages metadata and file system namespace
C. Executes Map tasks
D. Manages data replication
Answer: B
5. What is the role of DataNode in Hadoop?
A. Stores metadata
B. Stores actual data blocks
C. Monitors MapReduce jobs
D. Executes SQL queries
Answer: B
6. What is the default block size in Hadoop 3.x?
A. 64 MB
B. 128 MB
C. 256 MB
D. 512 MB
Answer: B
7. Which of the following is a resource management layer in Hadoop?
A. YARN
B. Pig
C. Sqoop
D. Oozie
Answer: A
8. Which language is primarily used for writing Hadoop MapReduce programs?
A. Python
B. Java
C. SQL
D. C++
Answer: B
9. What does MapReduce consist of?
A. Mapper and Combiner
B. Mapper and Reducer
C. Mapper only
D. Reducer only
Answer: B
10. What is a “combiner” in Hadoop MapReduce?
A. A backup reducer
B. A pre-reducer that performs local aggregation
C. A secondary mapper
D. A task scheduler
Answer: B
11. What does YARN stand for?
A. Yet Another Recursive NameNode
B. Yet Another Resource Negotiator
C. Your Advanced Resource Network
D. Yearly Assigned Resource Node
Answer: B
12. Which tool is used for data ingestion from RDBMS to Hadoop?
A. Pig
B. Hive
C. Sqoop
D. Flume
Answer: C
13. Which Hadoop ecosystem tool is used for real-time data ingestion?
A. Hive
B. Oozie
C. Flume
D. Sqoop
Answer: C
14. What type of language is Pig Latin in Hadoop?
A. Declarative
B. Procedural
C. Object-oriented
D. Functional
Answer: B
15. What is Hive mainly used for?
A. Workflow scheduling
B. Streaming analytics
C. Data warehousing and SQL-like queries
D. Data ingestion
Answer: C
16. What file format is best for columnar storage in Hadoop?
A. CSV
B. JSON
C. ORC
D. TXT
Answer: C
17. Which command is used to copy data from local to HDFS?
A. hadoop fs -get
B. hadoop fs -put
C. hadoop fs -delete
D. hadoop fs -ls
Answer: B
18. What happens if a DataNode fails in Hadoop?
A. The system stops
B. NameNode replaces it automatically
C. Data is replicated from another DataNode
D. All data is lost
Answer: C
19. Which of the following is a scheduling component in Hadoop?
A. ResourceManager
B. NodeManager
C. DataNode
D. NameNode
Answer: A
20. What is the major advantage of Hadoop?
A. High licensing cost
B. Centralized storage
C. Scalability and fault tolerance
D. Requires single-node architecture
























