I want to transfer files out from HDFS to local filesystem of a different server which is not in hadoop cluster but in the network.
I could have done:
hadoop fs -copyToLocal <src> <dest>
and then scp/ftp <toMyFileServer>.
As the data is huge and due to limited space on local filesystem of hadoop gateway machine, I wanted to avoid this and sent data directly to my file server.
Please help with some pointers on how to handle this issue.
This is the simplest way to do it:
ssh <YOUR_HADOOP_GATEWAY> "hdfs dfs -cat <src_in_HDFS> " > <local_dst>
It works for binary files too.