/** * Sets table_ based on targetTableName_ and performs table-type specific analysis: - Partition * clause is invalid for unpartitioned Hdfs tables and HBase tables - Overwrite is invalid for * HBase tables - Check INSERT privileges as well as write access to Hdfs paths - Cannot insert * into a view Adds table_ to the analyzer's descriptor table if analysis succeeds. */ private void setTargetTable(Analyzer analyzer) throws AnalysisException { // If the table has not yet been set, load it from the Catalog. This allows for // callers to set a table to analyze that may not actually be created in the Catalog. // One example use case is CREATE TABLE AS SELECT which must run analysis on the // INSERT before the table has actually been created. if (table_ == null) { if (!targetTableName_.isFullyQualified()) { targetTableName_ = new TableName(analyzer.getDefaultDb(), targetTableName_.getTbl()); } table_ = analyzer.getTable(targetTableName_, Privilege.INSERT); } else { targetTableName_ = new TableName(table_.getDb().getName(), table_.getName()); PrivilegeRequestBuilder pb = new PrivilegeRequestBuilder(); analyzer.registerPrivReq( pb.onTable(table_.getDb().getName(), table_.getName()) .allOf(Privilege.INSERT) .toRequest()); } // We do not support inserting into views. if (table_ instanceof View) { throw new AnalysisException( String.format("Impala does not support inserting into views: %s", table_.getFullName())); } boolean isHBaseTable = (table_ instanceof HBaseTable); int numClusteringCols = isHBaseTable ? 0 : table_.getNumClusteringCols(); if (partitionKeyValues_ != null && numClusteringCols == 0) { if (isHBaseTable) { throw new AnalysisException( "PARTITION clause is not valid for INSERT into " + "HBase tables. '" + targetTableName_ + "' is an HBase table"); } else { // Unpartitioned table, but INSERT has PARTITION clause throw new AnalysisException( "PARTITION clause is only valid for INSERT into " + "partitioned table. '" + targetTableName_ + "' is not partitioned"); } } if (table_ instanceof HdfsTable) { HdfsTable hdfsTable = (HdfsTable) table_; if (!hdfsTable.hasWriteAccess()) { throw new AnalysisException( String.format( "Unable to INSERT into target table " + "(%s) because Impala does not have WRITE access to at least one HDFS path" + ": %s", targetTableName_, hdfsTable.getFirstLocationWithoutWriteAccess())); } for (int colIdx = 0; colIdx < numClusteringCols; ++colIdx) { Column col = hdfsTable.getColumns().get(colIdx); // Hive has a number of issues handling BOOLEAN partition columns (see HIVE-6590). // Instead of working around the Hive bugs, INSERT is disabled for BOOLEAN // partitions in Impala. Once the Hive JIRA is resolved, we can remove this // analysis check. if (col.getType() == Type.BOOLEAN) { throw new AnalysisException( String.format( "INSERT into table with BOOLEAN " + "partition column (%s) is not supported: %s", col.getName(), targetTableName_)); } } } if (isHBaseTable && overwrite_) { throw new AnalysisException("HBase doesn't have a way to perform INSERT OVERWRITE"); } // Add target table to descriptor table. analyzer.getDescTbl().addReferencedTable(table_); }