mirror of
https://github.com/apache/impala.git
synced 2025-12-19 18:12:08 -05:00
This patch attempt to stabilize TestFetch by using HS2 as test protocol. test_rows_sent_counters is modified to use the default hs2_client. test_client_fetch_time_stats and test_client_fetch_time_stats_incomplete is modified to use MinimalHS2Connection that has more simpler mechanism in terms of fetching (ImpylaHS2Connection always fetch 10240 rows at a time). Implemented minimal functions needed to wait for finished state and pull runtime profile at MinimalHS2Connection. Testing: Loop the test 50 times and pass them all. Change-Id: I52651df37a318357711d26d2414e025cce4185c3 Reviewed-on: http://gerrit.cloudera.org:8080/22847 Reviewed-by: Impala Public Jenkins <impala-public-jenkins@cloudera.com> Tested-by: Impala Public Jenkins <impala-public-jenkins@cloudera.com>
234 lines
11 KiB
Python
234 lines
11 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
from __future__ import absolute_import, division, print_function
|
|
import re
|
|
|
|
from time import sleep
|
|
from tests.common.impala_connection import FINISHED, MinimalHS2Connection
|
|
from tests.common.impala_test_suite import IMPALAD_HS2_HOST_PORT, ImpalaTestSuite
|
|
from tests.common.test_dimensions import (
|
|
create_uncompressed_text_dimension,
|
|
extend_exec_option_dimension,
|
|
)
|
|
from tests.util.parse_util import parse_duration_string_ms, \
|
|
parse_duration_string_ns, get_time_summary_stats_counter
|
|
|
|
|
|
class TestFetch(ImpalaTestSuite):
|
|
"""Tests that are independent of whether result spooling is enabled or not."""
|
|
|
|
@classmethod
|
|
def add_test_dimensions(cls):
|
|
super(TestFetch, cls).add_test_dimensions()
|
|
# Result fetching should be independent of file format, so only test against
|
|
# text files.
|
|
cls.ImpalaTestMatrix.add_dimension(
|
|
create_uncompressed_text_dimension(cls.get_workload()))
|
|
|
|
def test_rows_sent_counters(self):
|
|
"""Validate that ClientFetchWaitTimer, NumRowsFetched, RowMaterializationRate,
|
|
and RowMaterializationTimer are set to valid values in the ImpalaServer section
|
|
of the runtime profile."""
|
|
num_rows = 25
|
|
query = "select sleep(100) from functional.alltypes limit {0}".format(num_rows)
|
|
client = self.hs2_client
|
|
handle = client.execute_async(query)
|
|
try:
|
|
# Wait until the query is 'FINISHED' and results are available for fetching.
|
|
client.wait_for_impala_state(handle, FINISHED, 30)
|
|
# Sleep for 2.5 seconds so that the ClientFetchWaitTimer is >= 1s.
|
|
sleep(2.5)
|
|
# Fetch the results so that the fetch related counters are updated.
|
|
assert client.fetch(query, handle).success
|
|
|
|
runtime_profile = client.get_runtime_profile(handle)
|
|
fetch_timer = re.search("ClientFetchWaitTimer: (.*)", runtime_profile)
|
|
assert fetch_timer and len(fetch_timer.groups()) == 1 and \
|
|
parse_duration_string_ms(fetch_timer.group(1)) > 1000
|
|
assert "NumRowsFetched: {0} ({0})".format(num_rows) in runtime_profile
|
|
assert re.search("RowMaterializationRate: [1-9]", runtime_profile)
|
|
# The query should take at least 1s to materialize all rows since it should sleep
|
|
# for at least 1s during materialization.
|
|
materialization_timer = re.search("RowMaterializationTimer: (.*)", runtime_profile)
|
|
assert materialization_timer and len(materialization_timer.groups()) == 1 and \
|
|
parse_duration_string_ms(materialization_timer.group(1)) > 1000
|
|
rpc_count = int(re.search("RPCCount: ([0-9]+)", runtime_profile).group(1))
|
|
assert rpc_count >= 5 and rpc_count <= 9
|
|
|
|
rpc_read_timer = re.search("RPCReadTimer: (.*)", runtime_profile)
|
|
assert rpc_read_timer and len(rpc_read_timer.groups()) == 1
|
|
rpc_read_ns = parse_duration_string_ns(rpc_read_timer.group(1))
|
|
assert 0 < rpc_read_ns and rpc_read_ns < 1000000
|
|
|
|
rpc_write_timer = re.search("RPCWriteTimer: (.*)", runtime_profile)
|
|
assert rpc_write_timer and len(rpc_write_timer.groups()) == 1
|
|
rpc_write_ns = parse_duration_string_ns(rpc_write_timer.group(1))
|
|
assert 0 < rpc_write_ns and rpc_write_ns < 10000000
|
|
|
|
create_result_time = re.search("CreateResultSetTime: (.*)", runtime_profile)
|
|
assert create_result_time and len(create_result_time.groups()) == 1
|
|
create_result_ms = parse_duration_string_ms(create_result_time.group(1))
|
|
assert 2400 < create_result_ms and create_result_ms < 2600
|
|
|
|
finally:
|
|
client.close_query(handle)
|
|
|
|
def test_client_fetch_time_stats(self):
|
|
num_rows = 27
|
|
client = MinimalHS2Connection(IMPALAD_HS2_HOST_PORT)
|
|
query = "select sleep(10) from functional.alltypes limit {0}".format(num_rows)
|
|
handle = client.execute_async(query)
|
|
try:
|
|
# Wait until the query is 'FINISHED' and results are available for fetching.
|
|
client.wait_for_finished_timeout(handle, 30)
|
|
|
|
# This loop will do 6 fetches that contain data and a final fetch with
|
|
# no data. The last fetch is after eos has been set, so it does not count.
|
|
rows_fetched = 0
|
|
while True:
|
|
result = client.fetch(query, handle, max_rows=5)
|
|
assert result is not None
|
|
rows_fetched += len(result)
|
|
# If no rows are returned, we are done.
|
|
if len(result) == 0:
|
|
break
|
|
sleep(0.1)
|
|
|
|
# After fetching all rows, sleep before closing the query. This should not
|
|
# count as client wait time, because the query is already done.
|
|
sleep(2.5)
|
|
finally:
|
|
client.close_query(handle)
|
|
|
|
runtime_profile = client.get_runtime_profile(handle)
|
|
|
|
summary_stats = get_time_summary_stats_counter("ClientFetchWaitTimeStats",
|
|
runtime_profile)
|
|
assert len(summary_stats) == 1
|
|
assert summary_stats[0].total_num_values == 6
|
|
# The 2.5 second sleep should not count, so the max must be less than 2.5 seconds.
|
|
assert summary_stats[0].max_value < 2500000000
|
|
assert summary_stats[0].min_value > 0
|
|
client.close()
|
|
|
|
def test_client_fetch_time_stats_incomplete(self):
|
|
num_rows = 27
|
|
client = MinimalHS2Connection(IMPALAD_HS2_HOST_PORT)
|
|
query = "select sleep(10) from functional.alltypes limit {0}".format(num_rows)
|
|
handle = client.execute_async(query)
|
|
try:
|
|
# Wait until the query is 'FINISHED' and results are available for fetching.
|
|
client.wait_for_finished_timeout(handle, 30)
|
|
|
|
# This loop will do 5 fetches for a total of 25 rows. This is incomplete.
|
|
for i in range(5):
|
|
result = client.fetch(query, handle, max_rows=5)
|
|
assert result is not None
|
|
sleep(0.1)
|
|
|
|
# Sleep before closing the query. For an incomplete fetch, this still counts
|
|
# towards the query time, so this does show up in the counters.
|
|
sleep(2.5)
|
|
finally:
|
|
client.close_query(handle)
|
|
|
|
runtime_profile = client.get_runtime_profile(handle)
|
|
|
|
summary_stats = get_time_summary_stats_counter("ClientFetchWaitTimeStats",
|
|
runtime_profile)
|
|
assert len(summary_stats) == 1
|
|
# There are 5 fetches and the finalization sample for a total of 6.
|
|
assert summary_stats[0].total_num_values == 6
|
|
# The 2.5 second sleep does count for an incomplete fetch, verify the max is higher.
|
|
assert summary_stats[0].max_value >= 2500000000
|
|
assert summary_stats[0].min_value > 0
|
|
client.close()
|
|
|
|
|
|
class TestFetchAndSpooling(ImpalaTestSuite):
|
|
"""Tests that apply when result spooling is enabled or disabled."""
|
|
|
|
@classmethod
|
|
def add_test_dimensions(cls):
|
|
super(TestFetchAndSpooling, cls).add_test_dimensions()
|
|
# Result fetching should be independent of file format, so only test against
|
|
# Parquet files.
|
|
cls.ImpalaTestMatrix.add_constraint(lambda v:
|
|
v.get_value('table_format').file_format == 'parquet')
|
|
# spool_query_results is set as true by default.
|
|
extend_exec_option_dimension(cls, 'spool_query_results', 'false')
|
|
|
|
def test_rows_sent_counters(self, vector):
|
|
"""Validate that RowsSent and RowsSentRate are set to valid values in
|
|
the PLAN_ROOT_SINK section of the runtime profile."""
|
|
num_rows = 10
|
|
if ('spool_query_results' in vector.get_value('exec_option')
|
|
and vector.get_value('exec_option')['spool_query_results'] == 'false'):
|
|
vector.get_value('exec_option')['debug_action'] = "BPRS_BEFORE_ADD_ROWS:SLEEP@1000"
|
|
else:
|
|
vector.get_value('exec_option')['debug_action'] = "BPRS_BEFORE_ADD_BATCH:SLEEP@1000"
|
|
result = self.execute_query("select id from functional.alltypes limit {0}"
|
|
.format(num_rows), vector.get_value('exec_option'))
|
|
assert "RowsSent: {0} ({0})".format(num_rows) in result.runtime_profile
|
|
rows_sent_rate = re.search(r"RowsSentRate: (\d*\.?\d*)", result.runtime_profile)
|
|
assert rows_sent_rate
|
|
assert float(rows_sent_rate.group(1)) > 0
|
|
|
|
|
|
class TestFetchTimeout(ImpalaTestSuite):
|
|
"""A few basic tests for FETCH_ROWS_TIMEOUT_MS that are not specific to the HS2 protocol
|
|
(e.g. in contrast to the tests in tests/hs2/test_fetch_timeout.py). These tests are
|
|
necessary because part of the FETCH_ROWS_TIMEOUT_MS code is HS2/Beeswax specific.
|
|
Unlike the tests in hs2/test_fetch_timeout.py, these tests do not validate that
|
|
individual RPC calls timeout, instead they set a low value for the timeout and assert
|
|
that the query works end-to-end."""
|
|
|
|
@classmethod
|
|
def add_test_dimensions(cls):
|
|
super(TestFetchTimeout, cls).add_test_dimensions()
|
|
# Result fetching should be independent of file format, so only test against
|
|
# Parquet files.
|
|
cls.ImpalaTestMatrix.add_constraint(lambda v:
|
|
v.get_value('table_format').file_format == 'parquet')
|
|
extend_exec_option_dimension(cls, 'spool_query_results', 'true')
|
|
|
|
def test_fetch_timeout(self, vector):
|
|
"""A simple test that runs a query with a low timeout and introduces delays in
|
|
RowBatch production. Asserts that the query succeeds and returns the expected number
|
|
of rows."""
|
|
num_rows = 100
|
|
query = "select * from functional.alltypes limit {0}".format(num_rows)
|
|
vector.get_value('exec_option')['batch_size'] = 1
|
|
vector.get_value('exec_option')['fetch_rows_timeout_ms'] = 1
|
|
vector.get_value('exec_option')['debug_action'] = '0:GETNEXT:DELAY'
|
|
results = self.execute_query(query, vector.get_value('exec_option'))
|
|
assert results.success
|
|
assert len(results.data) == num_rows
|
|
|
|
def test_fetch_before_finished_timeout(self, vector):
|
|
"""Tests that the FETCH_ROWS_TIMEOUT_MS timeout applies to queries that are not in
|
|
the 'finished' state. Similar to the test tests/hs2/test_fetch_timeout.py::
|
|
TestFetchTimeout::test_fetch_before_finished_timeout(_with_result_spooling)."""
|
|
num_rows = 10
|
|
query = "select * from functional.alltypes limit {0}".format(num_rows)
|
|
vector.get_value('exec_option')['debug_action'] = 'CRS_BEFORE_COORD_STARTS:SLEEP@5000'
|
|
vector.get_value('exec_option')['fetch_rows_timeout_ms'] = '1000'
|
|
results = self.execute_query(query, vector.get_value('exec_option'))
|
|
assert results.success
|
|
assert len(results.data) == num_rows
|