1. What is Hadoop Framework?
2. On What concepts the Hadoop Framework works?
3. What is MapReduce?
4. What is compute and storage modes?
5. How does master slave architecture in the Hadoop?
6. How does Hadoop application look like or their basic components?
7. Explain how input and output data format of the hadoop framework?
8. What are the restriction to the key and value class?
9. Explain the Word Count implementation via Hadoop framework?
10. Which intervace needs to be implemented to create Mapper and Reducer for the hadoop?
11. What Mapper does?
12. What is the InputSplit in map reduce software?
13. What is the InputFormat?
14. Where do you specify the Mapper implementation?
15. How Mapper is instantiated in a running Job?
16. Which are the methods in the Mapper interface?
17. What happens if you don't override the Mapper methods and keep them as it is?
18. What is the use of Context object?
19. How can u add the arbitrary key value pairs in your mapper?
20. How does Mapper's run() method works?
21. Which object can be used to get the progress of a particular job?
22.What is next step after Mapper or MapTask?
23. How can we control particular key should go in a specific reducer?
24. What is the use of Combiner?
25. How many maps are there in a particular job?
26. What is the Reducer used for?
27. Explain the core methods of the Reducer?
28. What are the primary phases of the Reducer?
29. Explain the shuffle?
30. Explain the Reducer Sort phase?
31. Explain the Reducer reduce phase?
32. How many Reducers should be configured?
33. It can be possible that a Job has 0 reducers?
34. What happens if number of reducers are 0?
35. How many instances of JobTracker can run on a Hadoop Cluser?
36. What is the JobTracker and what it performs in a Hadoop Cluster?
37. How a task is scheduled by a JobTracker?
38. How many instances of Tasktracker run on a Hadoop cluster?
39. What are the two main parts of the Hadoop framework?
40. Explain the use of TaskTracker in the Hadoop cluster?
41. What do you mean by TaskInstance?
42. How many daemon processes run on a Hadoop cluster?
43. How many maximum JVM can run on a slave node?
44. What is NAS?
45. How HDFA differs with NFS?
46. How does a NameNode handle the failure of the data nodes?
46. How does a NameNode handle the failure of the data nodes?
47. Can Reducer talk with each other?
48. Where the Mapper Intermediate data will be stored?
49. What is the use of Combiners in the Hadoop framework?
50. What is the Hadoop MapReduce API contract for a key and value Class?
51. What is a IdentityMapper and IdentityReducer in MapReduce?
52. What is the meaning of speculative execution in Hadoop? Why is it important?
53. When the reducers are are started in a MapReduce job?
54. What is HDFS ? How it is different from traditional file systems?
55. What is HDFS Block size? How is it different from traditional file system block size?
56.What is the full form of fsck?
57. What is a NameNode? How many instances of NameNode run on a Hadoop Cluster?
58. What is a DataNode? How many instances of DataNode run on a Hadoop Cluster?
59. How the Client communicates with HDFS?
60. How the HDFS Blocks are replicated?
2. On What concepts the Hadoop Framework works?
3. What is MapReduce?
4. What is compute and storage modes?
5. How does master slave architecture in the Hadoop?
6. How does Hadoop application look like or their basic components?
7. Explain how input and output data format of the hadoop framework?
8. What are the restriction to the key and value class?
9. Explain the Word Count implementation via Hadoop framework?
10. Which intervace needs to be implemented to create Mapper and Reducer for the hadoop?
11. What Mapper does?
12. What is the InputSplit in map reduce software?
13. What is the InputFormat?
14. Where do you specify the Mapper implementation?
15. How Mapper is instantiated in a running Job?
16. Which are the methods in the Mapper interface?
17. What happens if you don't override the Mapper methods and keep them as it is?
18. What is the use of Context object?
19. How can u add the arbitrary key value pairs in your mapper?
20. How does Mapper's run() method works?
21. Which object can be used to get the progress of a particular job?
22.What is next step after Mapper or MapTask?
23. How can we control particular key should go in a specific reducer?
24. What is the use of Combiner?
25. How many maps are there in a particular job?
26. What is the Reducer used for?
27. Explain the core methods of the Reducer?
28. What are the primary phases of the Reducer?
29. Explain the shuffle?
30. Explain the Reducer Sort phase?
31. Explain the Reducer reduce phase?
32. How many Reducers should be configured?
33. It can be possible that a Job has 0 reducers?
34. What happens if number of reducers are 0?
35. How many instances of JobTracker can run on a Hadoop Cluser?
36. What is the JobTracker and what it performs in a Hadoop Cluster?
37. How a task is scheduled by a JobTracker?
38. How many instances of Tasktracker run on a Hadoop cluster?
39. What are the two main parts of the Hadoop framework?
40. Explain the use of TaskTracker in the Hadoop cluster?
41. What do you mean by TaskInstance?
42. How many daemon processes run on a Hadoop cluster?
43. How many maximum JVM can run on a slave node?
44. What is NAS?
45. How HDFA differs with NFS?
46. How does a NameNode handle the failure of the data nodes?
46. How does a NameNode handle the failure of the data nodes?
47. Can Reducer talk with each other?
48. Where the Mapper Intermediate data will be stored?
49. What is the use of Combiners in the Hadoop framework?
50. What is the Hadoop MapReduce API contract for a key and value Class?
51. What is a IdentityMapper and IdentityReducer in MapReduce?
52. What is the meaning of speculative execution in Hadoop? Why is it important?
53. When the reducers are are started in a MapReduce job?
54. What is HDFS ? How it is different from traditional file systems?
55. What is HDFS Block size? How is it different from traditional file system block size?
56.What is the full form of fsck?
57. What is a NameNode? How many instances of NameNode run on a Hadoop Cluster?
58. What is a DataNode? How many instances of DataNode run on a Hadoop Cluster?
59. How the Client communicates with HDFS?
60. How the HDFS Blocks are replicated?