This document discusses challenges faced with running Hive at large scale at Yahoo. It describes how Yahoo runs Hive on 18 Hadoop clusters with over 400,000 nodes and 580PB of data. Even with optimizations like Tez, ORC, and vectorization, Yahoo encountered slow queries, out of memory errors, and slow partition pruning for queries on tables with millions of partitions. Fixes involved throwing more hardware at the metastore, client-side tuning, and addressing memory leaks and inefficiencies in the metastore and filesystem cache.