Complete the code to read a batch file from HDFS.
hadoop fs -cat /data/batch/[1]The batch ingestion reads static files like 'input.txt' from HDFS.
Complete the code to start a real-time data stream using Flume.
flume-ng agent --conf conf --conf-file [1] --name agent1 -Dflume.root.logger=INFO,consoleReal-time ingestion uses a streaming configuration file like 'stream.conf' for Flume.
Fix the error in the command to run a batch MapReduce job.
hadoop jar wordcount.jar WordCount [1] /outputBatch jobs read from batch input directories like '/input/batch'.
Fill both blanks to create a dictionary comprehension that filters batch files larger than 100MB.
large_files = {file: size for file, size in files.items() if size [1] [2]The comprehension filters files with size greater than 100MB.
Fill all three blanks to create a dictionary of real-time streams with their status if active.
active_streams = {stream[1]: status for stream, status in streams.items() if status [2] 'active' and stream.startswith([3])}The code filters streams with status 'active' and names starting with 'rt_', converting stream names to uppercase as keys.