==== ---- QUERY # Set an option explicitly; the test infrastructure should clear it before the next test. # The next test tests that buffer_pool_limit is unset (""). set buffer_pool_limit=7; ==== ---- QUERY set all; ---- RESULTS: VERIFY_IS_SUBSET 'ABORT_ON_DEFAULT_LIMIT_EXCEEDED','0','DEPRECATED' 'ABORT_ON_ERROR','0','REGULAR' 'ALLOW_UNSUPPORTED_FORMATS','0','DEPRECATED' 'BATCH_SIZE','0','DEVELOPMENT' 'BUFFER_POOL_LIMIT','','ADVANCED' 'DEBUG_ACTION','','DEVELOPMENT' 'DEFAULT_ORDER_BY_LIMIT','-1','DEPRECATED' 'DISABLE_CACHED_READS','0','DEPRECATED' 'DISABLE_CODEGEN','0','REGULAR' 'DISABLE_OUTERMOST_TOPN','0','DEVELOPMENT' 'EXPLAIN_LEVEL','1','REGULAR' 'HBASE_CACHE_BLOCKS','0','ADVANCED' 'HBASE_CACHING','0','ADVANCED' 'MAX_ERRORS','100','ADVANCED' 'MAX_IO_BUFFERS','0','DEPRECATED' 'MAX_SCAN_RANGE_LENGTH','0','DEVELOPMENT' 'MEM_LIMIT','0','REGULAR' 'NUM_NODES','0','DEVELOPMENT' 'NUM_SCANNER_THREADS','0','REGULAR' 'COMPRESSION_CODEC','','REGULAR' 'PARQUET_FILE_SIZE','0','ADVANCED' 'REQUEST_POOL','','REGULAR' 'RESERVATION_REQUEST_TIMEOUT','','DEPRECATED' 'RM_INITIAL_MEM','0','DEPRECATED' 'SYNC_DDL','0','REGULAR' 'V_CPU_CORES','','DEPRECATED' ---- TYPES STRING, STRING, STRING ==== ---- QUERY set explain_level=3; set all; ---- RESULTS: VERIFY_IS_SUBSET 'ABORT_ON_DEFAULT_LIMIT_EXCEEDED','0','DEPRECATED' 'ABORT_ON_ERROR','0','REGULAR' 'ALLOW_UNSUPPORTED_FORMATS','0','DEPRECATED' 'BATCH_SIZE','0','DEVELOPMENT' 'BUFFER_POOL_LIMIT','','ADVANCED' 'DEBUG_ACTION','','DEVELOPMENT' 'DEFAULT_ORDER_BY_LIMIT','-1','DEPRECATED' 'DISABLE_CACHED_READS','0','DEPRECATED' 'DISABLE_CODEGEN','0','REGULAR' 'DISABLE_OUTERMOST_TOPN','0','DEVELOPMENT' 'EXPLAIN_LEVEL','3','REGULAR' 'HBASE_CACHE_BLOCKS','0','ADVANCED' 'HBASE_CACHING','0','ADVANCED' 'MAX_ERRORS','100','ADVANCED' 'MAX_IO_BUFFERS','0','DEPRECATED' 'MAX_SCAN_RANGE_LENGTH','0','DEVELOPMENT' 'MEM_LIMIT','0','REGULAR' 'NUM_NODES','0','DEVELOPMENT' 'NUM_SCANNER_THREADS','0','REGULAR' 'COMPRESSION_CODEC','','REGULAR' 'PARQUET_FILE_SIZE','0','ADVANCED' 'REQUEST_POOL','','REGULAR' 'RESERVATION_REQUEST_TIMEOUT','','DEPRECATED' 'RM_INITIAL_MEM','0','DEPRECATED' 'SYNC_DDL','0','REGULAR' 'V_CPU_CORES','','DEPRECATED' ---- TYPES STRING, STRING, STRING ==== ---- QUERY set explain_level='0'; set all; ---- RESULTS: VERIFY_IS_SUBSET 'ABORT_ON_DEFAULT_LIMIT_EXCEEDED','0','DEPRECATED' 'ABORT_ON_ERROR','0','REGULAR' 'ALLOW_UNSUPPORTED_FORMATS','0','DEPRECATED' 'BATCH_SIZE','0','DEVELOPMENT' 'BUFFER_POOL_LIMIT','','ADVANCED' 'DEBUG_ACTION','','DEVELOPMENT' 'DEFAULT_ORDER_BY_LIMIT','-1','DEPRECATED' 'DISABLE_CACHED_READS','0','DEPRECATED' 'DISABLE_CODEGEN','0','REGULAR' 'DISABLE_OUTERMOST_TOPN','0','DEVELOPMENT' 'EXPLAIN_LEVEL','0','REGULAR' 'HBASE_CACHE_BLOCKS','0','ADVANCED' 'HBASE_CACHING','0','ADVANCED' 'MAX_ERRORS','100','ADVANCED' 'MAX_IO_BUFFERS','0','DEPRECATED' 'MAX_SCAN_RANGE_LENGTH','0','DEVELOPMENT' 'MEM_LIMIT','0','REGULAR' 'NUM_NODES','0','DEVELOPMENT' 'NUM_SCANNER_THREADS','0','REGULAR' 'COMPRESSION_CODEC','','REGULAR' 'PARQUET_FILE_SIZE','0','ADVANCED' 'REQUEST_POOL','','REGULAR' 'RESERVATION_REQUEST_TIMEOUT','','DEPRECATED' 'RM_INITIAL_MEM','0','DEPRECATED' 'SYNC_DDL','0','REGULAR' 'V_CPU_CORES','','DEPRECATED' ---- TYPES STRING, STRING, STRING ==== ---- QUERY set parquet_file_size='2g' ---- CATCH The PARQUET_FILE_SIZE query option must be less than 2GB. ==== ---- QUERY set foo=bar ---- CATCH Invalid query option: foo ==== ---- QUERY set parquet_compression_codec=bar ---- CATCH Invalid compression codec: bar ==== ---- QUERY # Test that SET actually does change the mem_limit. # First, show mem_limit is not hit. select 1 ---- RESULTS 1 ==== ---- QUERY # Set mem_limit really small so that queries will fail. set mem_limit=1; select count(string_col) from functional.alltypestiny ---- CATCH Memory limit exceeded ==== ---- QUERY # Set mem_limit back to unlimited and query should succeed again. set mem_limit=0; select count(string_col) from functional.alltypestiny ---- RESULTS 8 ---- TYPES BIGINT ==== ---- QUERY # IMPALA-3334: 'optimize_partition_key_scans' is a boolean query option set explain_level=0; set optimize_partition_key_scans=true; explain select min(month), max(year), ndv(day) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '01:AGGREGATE [FINALIZE]' '00:UNION' ' constant-operands=11' ==== ---- QUERY set explain_level=0; set optimize_partition_key_scans=1; explain select min(month), max(year), ndv(day) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '01:AGGREGATE [FINALIZE]' '00:UNION' ' constant-operands=11' ==== ---- QUERY set explain_level=0; set optimize_partition_key_scans=false; explain select min(month), max(year), ndv(day) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '03:AGGREGATE [FINALIZE]' '02:EXCHANGE [UNPARTITIONED]' '01:AGGREGATE' '00:SCAN HDFS [functional.alltypesagg]' ==== ---- QUERY set explain_level=0; set optimize_partition_key_scans=0; explain select min(month), max(year), ndv(day) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '03:AGGREGATE [FINALIZE]' '02:EXCHANGE [UNPARTITIONED]' '01:AGGREGATE' '00:SCAN HDFS [functional.alltypesagg]' ==== ---- QUERY set explain_level=0; set disable_streaming_preaggregations=false; explain select count(distinct double_col) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '06:AGGREGATE [FINALIZE]' '05:EXCHANGE [UNPARTITIONED]' '02:AGGREGATE' '04:AGGREGATE' '03:EXCHANGE [HASH(double_col)]' '01:AGGREGATE [STREAMING]' '00:SCAN HDFS [functional.alltypesagg]' ==== ---- QUERY set explain_level=0; set disable_streaming_preaggregations=0; explain select count(distinct double_col) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '06:AGGREGATE [FINALIZE]' '05:EXCHANGE [UNPARTITIONED]' '02:AGGREGATE' '04:AGGREGATE' '03:EXCHANGE [HASH(double_col)]' '01:AGGREGATE [STREAMING]' '00:SCAN HDFS [functional.alltypesagg]' ==== ---- QUERY set explain_level=0; set disable_streaming_preaggregations=true; explain select count(distinct double_col) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '06:AGGREGATE [FINALIZE]' '05:EXCHANGE [UNPARTITIONED]' '02:AGGREGATE' '04:AGGREGATE' '03:EXCHANGE [HASH(double_col)]' '01:AGGREGATE' '00:SCAN HDFS [functional.alltypesagg]' ==== ---- QUERY set explain_level=0; set disable_streaming_preaggregations=1; explain select count(distinct double_col) from functional.alltypesagg; ---- RESULTS: VERIFY_IS_SUBSET '06:AGGREGATE [FINALIZE]' '05:EXCHANGE [UNPARTITIONED]' '02:AGGREGATE' '04:AGGREGATE' '03:EXCHANGE [HASH(double_col)]' '01:AGGREGATE' '00:SCAN HDFS [functional.alltypesagg]' ==== ---- QUERY set max_row_size=-1; ---- CATCH Max row size must be a positive number of bytes: -1 ==== ---- QUERY set max_row_size=0; ---- CATCH Max row size must be a positive number of bytes: 0 ====