I used this Hive query to export a table into a CSV file.
INSERT OVERWRITE DIRECTORY '/user/data/output/test' select column1, column2 from table1;
The file generated '000000_0' does not have comma separator
Is this the right way to generate CSV file? If no, please let me know how can I generate the CSV file?
or use this
hive -e 'select * from your_Table' | sed 's/[\t]/,/g' > /home/yourfile.csv
You can also specify property set hive.cli.print.header=true
before the SELECT
to ensure that header along with data is created and copied to file.
For example:
hive -e 'set hive.cli.print.header=true; select * from your_Table' | sed 's/[\t]/,/g' > /home/yourfile.csv
If you don't want to write to local file system, pipe the output of sed
command back into HDFS
using the hadoop fs -put
command.
It may also be convenient to SFTP to your files using something like Cyberduck, or you can use scp
to connect via terminal / command prompt.