title | description | keywords | services | documentationcenter | author | manager | editor | ms.service | ms.workload | ms.tgt_pltfrm | ms.devlang | ms.topic | ms.date | ms.author |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Why do some joins in Hive give an Out of Memory error? | Microsoft Docs |
Use the Hive FAQ for answers to common questions on Hive on Azure HDInsight platform. |
Azure HDInsight, ambari, Tez, FAQ, troubleshooting guide, |
Azure HDInsight |
na |
multiple |
na |
na |
na |
article |
01/17/2018 |
msft-tacox |
The default behavior for Hive joins is to load the entire contents of a table into memory so that a join can be performed without having to perform a Map/Reduce step. If the Hive table is too large to fit into memory, the query can fail.
When running joins in hive of sufficient size, the following error is encountered:
Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded error.
Prevent Hive from loading tables into memory on joins (instead performing a Map/Reduce step) by setting the following Hive configuration value:
hive.auto.convert.join=false