Month: <span>July 2016</span>

Month: July 2016

Yarn Job Failed with Error: “Split metadata size exceeded 10000000”

When you run a really big job in Hive that failed with the following error: This indicated that the value for mapreduce.job.split.metainfo.maxsize is too small for your job (default value of 10000000). There are two options to fix this: 1. Set the value of mapreduce.job.split.metainfo.maxsize to be “-1” (unlimited) specifically …

My new Snowflake Blog is now live. I will not be updating this blog anymore but will continue with new contents in the Snowflake world!