A customer has an Isilon cluster that they want to simplify a Hadoop workflow. The workflow analyzes large amounts of log data written to storage using FTP and results are viewed by Microsoft Windows clients?
How can this be achieved?
A . FTP, SMB, and HDFS NameNode and DataNode protocol support on the same file system to enable each workflow step to access data from the same location avoiding data migration.
B . SyncIQ to migrate the log data between an Isilon cluster and another Hadoop cluster, to retrieve results from the Hadoop cluster, and to store them in an SMB share.
C . SmartConnect to direct clients to an external Hadoop NameNode and to SMB shares so data ingest, analytics, and results phases are transparently directed.
D . FTP and SMB protocol support to provide log ingest and Windows clients; SmartPools will stub to HDFS helping to reduce the frequency of external data migration.
Answer: A
Leave a Reply