These Hadoop multiple-choice questions and their answers will help you strengthen your grip on the subject of Hadoop You can prepare for an upcoming exam or job interview with these 100+ Hadoop MCQs.
So scroll down and start answering.
A. setReplication(string replication, Path p, FsPermission permission)
B. setReplication(FsPermission permission, short replication)
C. setReplication(string replication, Path p)
D. setReplication(Path src, short replication)
A. Only i) and ii)
B. Only ii) and iii)
C. Only i) and iv)
D. Only iii) and iv)
E. Only i), ii), and iii)
A. Tasktracker
B. Jobtracker
C. YARN
D. Node Manager
A. YARN_CONF_DIR
B. HADOOP_PREFIX
C. HADOOP_CONF_DIR
D. HADOOP_HOME
A. fetchdt
B. dfs
C. oiv
D. fsck
A. bin/hdfs dfs −cat /newexample/example1.txt
B. bin/hadoop dfsadmin -ddir /newexample/example1.txt
C. bin/hadoop dfs -cat /newexample/example1.txt
D. bin/hdfs dfsadmin -ddir /newexample/example1.txt
A. Resource Manager
B. Node Manager
C. Application Master
D. Container
A. Transform
B. Command Line
C. ObjectInspector
D. Thrift
A. It displays the summary of file lengths.
B. In case of a file, it displays the length of the file, while in case of a directory, it displays the sizes of the files and directories present in that directory.
C. It displays the number of files in a particular directory.
D. It displays the numbers and names of files present in a particular directory.
A. Only i) and iii)
B. Only ii) and iv)
C. Only i), ii), and iii)
D. Only i), iii), and iv)
E. All i), ii), iii), and iv)
A. mapreduce.task.io.sort.mb
B. io.record.sort.percent
C. mapreduce.partitioner.class
D. mapreduce.task.io.mb
A. PigReducerEstimator interface
B. StoreFunc interface
C. FilterFunc interface
D. Accumulator interface
A. ORDER BY
B. SORT BY
C. Either a or b
D. None of the above
A. CREATE MyData DATABASE
B. CREATE DATABASE MyData
C. CREATE NEW MyData DATABASE
D. CREATE NEW DATABASE MyData
A. It is used for checking whether all the libraries are available.
B. It is used for expanding the wildcards.
C. It is used for specifying a resource manager.
D. It is used for assigning a value to a property
A. Interactive mode
B. Batch mode
C. Embedded mode
D. Either Interactive or Batch mode
A. bytearray
B. Bag
C. Map
D. Tuple
A. It provides insights on the status of the application.
B. It guarantees the restart on the application and hardware failures.
C. It allocates resources to the applications running in the cluster.
D. It handles the applications submitted by the clients.
A. It does not impose any restrictions on the schema of the two datasets that are being concatenated.
B. It removes the duplicate tuples while concatenating the datasets.
C. It preserves the ordering of the tuples while concatenating the datasets.
D. It uses the ONSCHEMA qualifier for giving a schema to the result
A. Cogroup
B. Foreach
C. Cross
D. Union
A. hadoop command [streamingOptions]
B. command ∼ Hadoop [genericOptions] [streamingOptions]
C. hadoop command [genericOptions] [streamingOptions]
D. command ∼ Hadoop [streamingOptions] [genericOptions]
A. The joins in Hive are commutative.
B. In Hive, more than two tables can be joined.
C. The first table participating in the join is streamed to the reduce task by default.
D. All are correct.
A. hdfs chgrp [owner] [:[group] ] [-R] <filepath><newgroup>
B. hdfs chgrp [-R] <group> <filepath>
C. hdfs chgrp [-R] <[group[: [owner]> <filepath>
D. hdfs chgrp <group> <filepath>[-R] <newgroup>
E. hdfs chgrp <group>[-R] <newgroup>
A. LevelDB based state-store
B. FileSystem based state-store
C. Zookeeper based state-store
D. Either option a or b could be used
A. User access permission is not implemented in HDFS.
B. In HDFS, a user is not allowed to create directories.
C. HDFS supports hard links.
D. HDFS implements user quotas.
A. touchz
B. tail
C. text
D. test
A. hadoop fs ∼ D blocksize=536870912 -put abc.dat.txt to abc.dat.newblock.txt
B. hadoop fs.blocksize=536870912 -put abc.dat.txt abc.dat.newblock.txt
C. hadoop fs -D dfs.blocksize=536870912 -put abc.dat.txt abc.dat.newblock.txt
D. hadoop fs.blocksize −D=536870912 -put abc.dat.txt to abc.dat.newblock.txt
A. It is used for identifying the class to be used by the client.
B. It used for listing the logical ID's used by Resource Managers.
C. It is used for specifying the corresponding host name for Resource Manager.
D. It is used for identifying the Resource Manager in ensemble.
A. The scaling in MapReduce is non-linear, whereas in RDBMS it is linear.
B. In MapReduce, the read operation can be performed many times but the write operation can be performed only once. In traditional RDBMS, both read and write operations can be performed many times.
C. The integrity of MapReduce is higher as compared to RDBMS.
D. The access pattern of MapReduce is Batch, whereas the access pattern of RDBMS is Interactive and Batch
A. It prevents the Application Manager from providing links to the malicious external sites.
B. It prevents the execution of the malicious JavaScript code.
C. It strips the cookies from the user and replaces them with a single cookie, providing the user name of the logged in user.
D. It runs as a part of Resource Manger but cannot be configured to run in the stand-alone mode.
A. −output directoryname
B. −cmdenv name = value
C. −combiner streamingCommand
D. −reducer JavaClassName
A. -file ABC =/home/example/
B. −mapper ABC = /home/inputReader/example/dictionaries/
C. −input ABC = /home/directories/example
D. -cmdenv ABC = /home/example/dictionaries/
A. It provides high performance.
B. It can be deployed easily.
C. It is highly flexible.
D. It can protect against software as well as physical threats.
A. hadoop fs −chmod abc 310
B. hadoop fs −chmod 740 abc
C. hadoop fs ∼chmod 420 abc
D. hadoop fs −chmod abc ∼ 860
A. Read
B. Write
C. Execute
D. All three permission levels are allowed
A. Left semi-join
B. Inner join
C. Theta join
D. Fuzzy join
A. kinit
B. klist
C. ktutil
D. mradmin
E. dfsadmin
A. yarn.timeline-service.ttl-enable
B. yarn.timeline-service.enabled
C. yarn.timeline-service.generic-application-history.enabled
D. yarn.timeline-service.recovery.enabled
A. yarn node -jar [main class name] <jar file path>[arguments…]
B. yarn jar <jar file path> [main class name] [arguments…]
C. yarn application -jar [main class name] <jar file path> [arguments…]
D. yarn logs jar <jar file path> [main class name] [arguments…]
A. --config confdir
B. --config. YarnConfiguration
C. daemonlog −getlevel
D. daemonlog confdir
A. set
B. set −v
C. dfs
D. reset
A. It is used for executing a dfs command from the Hive shell.
B. It is used for executing a shell command inside the CLI.
C. It is used for executing a shell command from the Hive shell.
D. It is used for executing a dfs command inside the CLI.
A. Only i) and iii)
B. Only i) and ii)
C. Only i), ii), and iv)
D. All i), ii), iii), and iv)
A. yarn.resourcemanager.ha.rm-ids
B. yarn.resourcemanager.address.rm-id
C. yarn.resourcemanager.hostname.rm-id
D. yarn.resourcemanager.scheduler.address.rm-id
A. setgid
B. setFile
C. setfattr
D. setQuota
E. setConf
A. DESCRIBE
B. ILLUSTRATE
C. EXPLAIN
D. Both a and b
A. $ mvn package − Pdocs
B. $ mvn Pdocs
C. $ curl − mvn Pdocs
D. $ curl − mvn Pdocs − package
A. security.client.datanode.protocol.acl
B. security.namenode.protocol.acl
C. security.client.protocol.acl
D. security.datanode.protocol.acl
A. 1 day
B. 3 days
C. 5 days
D. 7 days
A. org.apache.hadoop.security.AuthenticationInitializer class name
B. org.apache.hadoop.security.ShellBasedUnixGroupsMapping class name
C. org.apache.hadoop.security.LdapGroupsMapping class
D. org.apache.hadoop.security.ssl class name