mirror of
https://github.com/apache/impala.git
synced 2026-01-04 09:00:56 -05:00
Various test tools and frameworks, including the stress test, random query generator, and nested types loader, share common modules. This change IMPALA-3980: qgen: re-enable Hive as a target database made changes to tests.comparison.cli_options, the shared command line option module, and to tests.comparison.cluster, the shared module for modeling various Impala clusters. Those changes were for the random query generator, but didn't take into account the other shared entry points. It was possible to call some of those entry points in such a way as to produce an exception, because the Hive-related options are now required for miniclusters, but the Hive-related options weren't always being initialized in those entry points. The simple fix is to say that, because Hive settings are now needed to create Minicluster objects, make the Hive options initialized with cluster options, not connection options. While I was making these changes, I fixed all flake8 problems in this file. Testing: - qgen/minicluster unit tests (regression test) - full private data load job, including load_nested.py (bug verification) - data_generator.py run (regression test), long enough to verify connection to the minicluster, using both Hive and Impala - discrepancy_searcher.py run (regression test), long enough verify connection to the minicluster, using both Hive and Impala - concurrent_select.py (in typical mode using a CM host, this is a regression check; from the command line against the minicluster, this is a bug verification) Change-Id: I2a2915e6db85ddb3d8e1bce8035eccd0c9324b4b Reviewed-on: http://gerrit.cloudera.org:8080/4555 Reviewed-by: Michael Brown <mikeb@cloudera.com> Reviewed-by: Ishaan Joshi <ishaan@cloudera.com> Tested-by: Internal Jenkins
292 lines
10 KiB
Python
292 lines
10 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
'''Helpers for parsing command line options'''
|
|
|
|
import logging
|
|
import os
|
|
import sys
|
|
from getpass import getuser
|
|
from tempfile import gettempdir
|
|
|
|
import db_connection
|
|
from cluster import (
|
|
CmCluster,
|
|
DEFAULT_HIVE_HOST,
|
|
DEFAULT_HIVE_PASSWORD,
|
|
DEFAULT_HIVE_PORT,
|
|
DEFAULT_HIVE_USER,
|
|
MiniCluster,
|
|
MiniHiveCluster,
|
|
)
|
|
from db_types import TYPES
|
|
|
|
|
|
def add_logging_options(parser, default_debug_log_file=None):
|
|
if not default_debug_log_file:
|
|
default_debug_log_file = os.path.join(
|
|
gettempdir(), os.path.basename(sys.modules["__main__"].__file__) + ".log")
|
|
parser.add_argument(
|
|
'--log-level', default='INFO',
|
|
help='The log level to use.', choices=('DEBUG', 'INFO', 'WARN', 'ERROR'))
|
|
parser.add_argument(
|
|
'--debug-log-file', default=default_debug_log_file,
|
|
help='Path to debug log file.')
|
|
|
|
|
|
def configure_logging(log_level, debug_log_file=None, log_thread_id=False,
|
|
log_process_id=False):
|
|
root_logger = logging.getLogger()
|
|
root_logger.setLevel(logging.DEBUG)
|
|
|
|
console_logger = logging.StreamHandler(sys.stdout)
|
|
console_logger.name = "console"
|
|
console_logger.setLevel(getattr(logging, log_level))
|
|
format = "%(asctime)s"
|
|
if log_process_id:
|
|
format += " %(process)d"
|
|
if log_thread_id:
|
|
format += " %(thread)d"
|
|
format += " %(levelname)s:%(module)s[%(lineno)s]:%(message)s"
|
|
console_logger.setFormatter(logging.Formatter(format, "%H:%M:%S"))
|
|
root_logger.addHandler(console_logger)
|
|
|
|
if debug_log_file:
|
|
file_logger = logging.FileHandler(debug_log_file, mode="w")
|
|
file_logger.name = "file"
|
|
file_logger.setFormatter(logging.Formatter(format, "%H:%M:%S"))
|
|
file_logger.setLevel(logging.DEBUG)
|
|
root_logger.addHandler(file_logger)
|
|
|
|
def create_third_party_filter(level):
|
|
def filter_record(record):
|
|
name = record.name
|
|
if name.startswith("impala.") or name.startswith("paramiko.") or \
|
|
name.startswith("hdfs") or name.startswith("requests"):
|
|
return record.levelno >= level
|
|
return True
|
|
log_filter = logging.Filter()
|
|
log_filter.filter = filter_record
|
|
return log_filter
|
|
console_logger.addFilter(create_third_party_filter(logging.WARN))
|
|
if debug_log_file:
|
|
file_logger.addFilter(create_third_party_filter(logging.INFO))
|
|
|
|
|
|
def add_ssh_options(parser):
|
|
parser.add_argument(
|
|
'--ssh-user', metavar='user name', default=getuser(),
|
|
help='The user name to use for SSH connections to cluster nodes.')
|
|
parser.add_argument(
|
|
'--ssh-key-file', metavar='path to file',
|
|
help='Specify an additional SSH key other than the defaults in ~/.ssh.')
|
|
parser.add_argument(
|
|
'--ssh-port', metavar='number', type=int, default=22,
|
|
help='The port number to use when connecting through SSH.')
|
|
|
|
|
|
def add_db_name_option(parser):
|
|
parser.add_argument(
|
|
'--db-name', default='randomness',
|
|
help='The name of the database to use. Ex: functional.')
|
|
|
|
|
|
def add_cluster_options(parser):
|
|
add_minicluster_options(parser)
|
|
add_cm_options(parser)
|
|
add_ssh_options(parser)
|
|
parser.add_argument(
|
|
'--hadoop-user-name', default=getuser(),
|
|
help='The user name to use when interacting with hadoop.')
|
|
|
|
|
|
def add_minicluster_options(parser):
|
|
group = parser.add_argument_group('Hive Options')
|
|
group.add_argument(
|
|
'--use-hive', action='store_true', default=False,
|
|
help='Use Hive (Impala will be skipped)')
|
|
group.add_argument(
|
|
'--hive-host', default=DEFAULT_HIVE_HOST,
|
|
help='The name of the host running the HS2')
|
|
group.add_argument(
|
|
'--hive-port', default=DEFAULT_HIVE_PORT, type=int,
|
|
help='The port of HiveServer2')
|
|
group.add_argument(
|
|
'--hive-user', default=DEFAULT_HIVE_USER,
|
|
help='The user name to use when connecting to HiveServer2')
|
|
group.add_argument(
|
|
'--hive-password', default=DEFAULT_HIVE_PASSWORD,
|
|
help='The password to use when connecting to HiveServer2')
|
|
parser.add_argument_group(group)
|
|
|
|
parser.add_argument(
|
|
'--minicluster-num-impalads', default=3, type=int, metavar='num impalads',
|
|
help='The number of impalads in the mini cluster.')
|
|
|
|
|
|
def add_cm_options(parser):
|
|
parser.add_argument(
|
|
'--cm-host', metavar='host name',
|
|
help='The host name of the CM server.')
|
|
parser.add_argument(
|
|
'--cm-port', default=7180, type=int, metavar='port number',
|
|
help='The port of the CM server.')
|
|
parser.add_argument(
|
|
'--cm-user', default="admin", metavar='user name',
|
|
help='The name of the CM user.')
|
|
parser.add_argument(
|
|
'--cm-password', default="admin", metavar='password',
|
|
help='The password for the CM user.')
|
|
parser.add_argument(
|
|
'--cm-cluster-name', metavar='name',
|
|
help='If CM manages multiple clusters, use this to specify which cluster to use.')
|
|
|
|
|
|
def create_cluster(args):
|
|
if args.cm_host:
|
|
cluster = CmCluster(
|
|
args.cm_host, user=args.cm_user, password=args.cm_password,
|
|
cluster_name=args.cm_cluster_name, ssh_user=args.ssh_user, ssh_port=args.ssh_port,
|
|
ssh_key_file=args.ssh_key_file)
|
|
elif args.use_hive:
|
|
cluster = MiniHiveCluster(args.hive_host, args.hive_port)
|
|
else:
|
|
cluster = MiniCluster(args.hive_host, args.hive_port, args.minicluster_num_impalads)
|
|
cluster.hadoop_user_name = args.hadoop_user_name
|
|
return cluster
|
|
|
|
|
|
def add_storage_format_options(parser):
|
|
storage_formats = ['avro', 'parquet', 'rcfile', 'sequencefile', 'textfile']
|
|
parser.add_argument(
|
|
'--storage-file-formats', default=','.join(storage_formats),
|
|
help='A comma separated list of storage formats to use.')
|
|
|
|
|
|
def add_data_types_options(parser):
|
|
parser.add_argument(
|
|
'--data-types', default=','.join(type_.__name__ for type_ in TYPES),
|
|
help='A comma separated list of data types to use.')
|
|
|
|
|
|
def add_timeout_option(parser):
|
|
parser.add_argument(
|
|
'--timeout', default=(3 * 60), type=int, help='Query timeout in seconds')
|
|
|
|
|
|
def add_connection_option_groups(parser):
|
|
|
|
group = parser.add_argument_group('MySQL Options')
|
|
group.add_argument(
|
|
'--use-mysql', action='store_true', help='Use MySQL')
|
|
group.add_argument(
|
|
'--mysql-host', default='localhost',
|
|
help='The name of the host running the MySQL database.')
|
|
group.add_argument(
|
|
'--mysql-port', default=3306, type=int,
|
|
help='The port of the host running the MySQL database.')
|
|
group.add_argument(
|
|
'--mysql-user', default='root',
|
|
help='The user name to use when connecting to the MySQL database.')
|
|
group.add_argument(
|
|
'--mysql-password',
|
|
help='The password to use when connecting to the MySQL database.')
|
|
parser.add_argument_group(group)
|
|
|
|
group = parser.add_argument_group('Oracle Options')
|
|
group.add_argument('--use-oracle', action='store_true', help='Use Oracle')
|
|
group.add_argument(
|
|
'--oracle-host', default='localhost',
|
|
help='The name of the host running the Oracle database.')
|
|
group.add_argument(
|
|
'--oracle-port', default=1521, type=int,
|
|
help='The port of the host running the Oracle database.')
|
|
group.add_argument(
|
|
'--oracle-user', default='system',
|
|
help='The user name to use when connecting to the Oracle database.')
|
|
group.add_argument(
|
|
'--oracle-password',
|
|
help='The password to use when connecting to the Oracle database.')
|
|
parser.add_argument_group(group)
|
|
|
|
group = parser.add_argument_group('Postgresql Options')
|
|
group.add_argument(
|
|
'--use-postgresql', action='store_true', help='Use Postgresql')
|
|
group.add_argument(
|
|
'--postgresql-host', default='localhost',
|
|
help='The name of the host running the Postgresql database.')
|
|
group.add_argument(
|
|
'--postgresql-port', default=5432, type=int,
|
|
help='The port of the host running the Postgresql database.')
|
|
group.add_argument(
|
|
'--postgresql-user', default='postgres',
|
|
help='The user name to use when connecting to the Postgresql database.')
|
|
group.add_argument(
|
|
'--postgresql-password',
|
|
help='The password to use when connecting to the Postgresql database.')
|
|
parser.add_argument_group(group)
|
|
|
|
|
|
def get_db_type(args):
|
|
db_types = list()
|
|
if args.use_mysql:
|
|
db_types.append(db_connection.MYSQL)
|
|
if args.use_oracle:
|
|
db_types.append(db_connection.ORACLE)
|
|
if args.use_postgresql:
|
|
db_types.append(db_connection.POSTGRESQL)
|
|
if not db_types:
|
|
raise Exception(
|
|
"At least one of --use-mysql, --use-oracle, or --use-postgresql must be used")
|
|
elif len(db_types) > 1:
|
|
raise Exception("Too many databases requested: %s" % db_types)
|
|
return db_types[0]
|
|
|
|
|
|
def create_connection(args, db_type=None, db_name=None):
|
|
if not db_type:
|
|
db_type = get_db_type(args)
|
|
if db_type == db_connection.POSTGRESQL:
|
|
conn_class = db_connection.PostgresqlConnection
|
|
elif db_type == db_connection.MYSQL:
|
|
conn_class = db_connection.MySQLConnection
|
|
elif db_type == db_connection.ORACLE:
|
|
conn_class = db_connection.OracleConnection
|
|
elif db_type == db_connection.HIVE:
|
|
conn_class = db_connection.HiveConnection
|
|
else:
|
|
raise Exception(
|
|
'Unexpected db_type: %s; expected one of %s.' % (
|
|
db_type, ', '.join([db_connection.POSTGRESQL, db_connection.MYSQL,
|
|
db_connection.ORACLE])))
|
|
prefix = db_type.lower()
|
|
return conn_class(
|
|
user_name=getattr(args, prefix + '_user'),
|
|
password=getattr(args, prefix + '_password'),
|
|
host_name=getattr(args, prefix + '_host'),
|
|
port=getattr(args, prefix + '_port'),
|
|
db_name=db_name)
|
|
|
|
|
|
def add_kerberos_options(parser):
|
|
parser.add_argument(
|
|
"--use-kerberos", action="store_true",
|
|
help="Use kerberos when communicating with Impala. This requires that kinit has"
|
|
" already been done before running this script.")
|
|
parser.add_argument(
|
|
"--kerberos-principal", default=getuser(), help="The principal name to use.")
|