Files
impala/tests/metadata/test_load.py
Casey Ching 074e5b4349 Remove hashbang from non-script python files
Many python files had a hashbang and the executable bit set though
they were not intended to be run a standalone script. That makes
determining which python files are actually scripts very difficult.
A future patch will update the hashbang in real python scripts so they
use $IMPALA_HOME/bin/impala-python.

Change-Id: I04eafdc73201feefe65b85817a00474e182ec2ba
Reviewed-on: http://gerrit.cloudera.org:8080/599
Reviewed-by: Casey Ching <casey@cloudera.com>
Reviewed-by: Taras Bobrovytsky <tbobrovytsky@cloudera.com>
Tested-by: Internal Jenkins
2015-08-04 05:26:07 +00:00

87 lines
3.5 KiB
Python

# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Functional tests for LOAD DATA statements.
import os
import pytest
from tests.common.impala_test_suite import ImpalaTestSuite
from tests.common.test_dimensions import (
create_single_exec_option_dimension,
create_uncompressed_text_dimension)
from tests.common.skip import SkipIfS3
from tests.util.filesystem_utils import WAREHOUSE
TEST_TBL_PART = "test_load"
TEST_TBL_NOPART = "test_load_nopart"
STAGING_PATH = 'test-warehouse/test_load_staging'
ALLTYPES_PATH = "test-warehouse/alltypes/year=2010/month=1/100101.txt"
MULTIAGG_PATH = 'test-warehouse/alltypesaggmultifiles/year=2010/month=1/day=1'
HIDDEN_FILES = ["{0}/3/.100101.txt".format(STAGING_PATH),
"{0}/3/_100101.txt".format(STAGING_PATH)]
@SkipIfS3.load_data
class TestLoadData(ImpalaTestSuite):
@classmethod
def get_workload(self):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestLoadData, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(create_single_exec_option_dimension())
cls.TestMatrix.add_dimension(create_uncompressed_text_dimension(cls.get_workload()))
def _clean_test_tables(self):
self.client.execute("drop table if exists functional.{0}".format(TEST_TBL_NOPART))
self.client.execute("drop table if exists functional.{0}".format(TEST_TBL_PART))
self.hdfs_client.delete_file_dir(STAGING_PATH, recursive=True)
def teardown_method(self, method):
self._clean_test_tables()
def setup_method(self, method):
# Defensively clean the data dirs if they exist.
self._clean_test_tables()
# Create staging directories for load data inpath. The staging directory is laid out
# as follows:
# - It has 6 sub directories, numbered 1-6
# - The directories are populated with files from a subset of partitions in existing
# partitioned tables.
# - Sub Directories 1-4 have single files copied from alltypes/
# - Sub Directories 5-6 have multiple files (4) copied from alltypesaggmultifiles
# - Sub Directory 3 also has hidden files, in both supported formats.
for i in xrange(1, 6):
self.hdfs_client.make_dir('{0}/{1}'.format(STAGING_PATH, i), permission=777)
# Copy single file partitions from alltypes.
for i in xrange(1, 4):
self.hdfs_client.copy(ALLTYPES_PATH, "{0}/{1}/100101.txt".format(STAGING_PATH, i))
# Copy multi file partitions from alltypesaggmultifiles.
file_infos = self.hdfs_client.list_dir(
MULTIAGG_PATH).get('FileStatuses').get('FileStatus')
file_names = [info.get('pathSuffix') for info in file_infos]
for i in xrange(4, 6):
for file_ in file_names:
self.hdfs_client.copy(
"{0}/{1}".format(MULTIAGG_PATH, file_),
'{0}/{1}/{2}'.format(STAGING_PATH, i, file_))
# Create two hidden files, with a leading . and _
for file_ in HIDDEN_FILES:
self.hdfs_client.copy(ALLTYPES_PATH, file_)
# Create both the test tables.
self.client.execute("create table functional.{0} like functional.alltypes"
" location '{1}/{0}'".format(TEST_TBL_PART, WAREHOUSE))
self.client.execute("create table functional.{0} like functional.alltypesnopart"
" location '{1}/{0}'".format(TEST_TBL_NOPART, WAREHOUSE))
def test_load(self, vector):
self.run_test_case('QueryTest/load', vector)
# The hidden files should not have been moved as part of the load operation.
for file_ in HIDDEN_FILES:
assert self.hdfs_client.get_file_dir_status(file_), "{0} does not exist".format(
file_)