Month: <span>July 2016</span>

Month: July 2016

Yarn Job Failed with Error: “Split metadata size exceeded 10000000”

When you run a really big job in Hive that failed with the following error: This indicated that the value for mapreduce.job.split.metainfo.maxsize is too small for your job (default value of 10000000). There are two options to fix this: 1. Set the value of mapreduce.job.split.metainfo.maxsize to be “-1” (unlimited) specifically …

Loading

Hive’s Staging Directory Not Cleaned Up Properly

This article explains the situation that will cause Hive to leave its staging directories that were created during processing and not doing clean up after job finished successfully. The issue happens when user runs Hive query through Hue’s Hive Editor, it does not apply to queries running through Beeline, Hive …

Loading

My new Snowflake Blog is now live. I will not be updating this blog anymore but will continue with new contents in the Snowflake world!