mirror of
https://github.com/apache/impala.git
synced 2026-01-07 00:02:28 -05:00
Increase the timeout to over 120s to match datastream_sender_timeout_ms. This should avoid spurious test failures if we are unlucky and a sender gets stuck waiting for a receiver fragment that will never start. Testing: Ran the test in a loop for a while to flush out any flakiness. Change-Id: I9fe6e6c74538d0747e3eeb578cf0518494ff10c8 Reviewed-on: http://gerrit.cloudera.org:8080/5244 Tested-by: Impala Public Jenkins Reviewed-by: Tim Armstrong <tarmstrong@cloudera.com>
70 lines
2.8 KiB
Python
70 lines
2.8 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import pytest
|
|
from tests.beeswax.impala_beeswax import ImpalaBeeswaxException
|
|
from tests.common.impala_test_suite import ImpalaTestSuite
|
|
from tests.common.impala_cluster import ImpalaCluster
|
|
from tests.verifiers.metric_verifier import MetricVerifier
|
|
|
|
class TestFragmentLifecycle(ImpalaTestSuite):
|
|
"""Using the debug action interface, check that failed queries correctly clean up *all*
|
|
fragments"""
|
|
|
|
IN_FLIGHT_FRAGMENTS = "impala-server.num-fragments-in-flight"
|
|
@classmethod
|
|
def get_workload(self):
|
|
return 'functional'
|
|
|
|
@pytest.mark.execute_serially
|
|
def test_failure_in_prepare(self):
|
|
# Fail the scan node
|
|
verifiers = [ MetricVerifier(i.service) for i in ImpalaCluster().impalads ]
|
|
self.client.execute("SET DEBUG_ACTION='-1:0:PREPARE:FAIL'");
|
|
try:
|
|
self.client.execute("SELECT COUNT(*) FROM functional.alltypes")
|
|
assert "Query should have thrown an error"
|
|
except ImpalaBeeswaxException:
|
|
pass
|
|
|
|
for v in verifiers:
|
|
v.wait_for_metric(self.IN_FLIGHT_FRAGMENTS, 0)
|
|
|
|
@pytest.mark.execute_serially
|
|
def test_failure_in_prepare_multi_fragment(self):
|
|
# Test that if one fragment fails that the others are cleaned up during the ensuing
|
|
# cancellation.
|
|
verifiers = [ MetricVerifier(i.service) for i in ImpalaCluster().impalads ]
|
|
# Fail the scan node
|
|
self.client.execute("SET DEBUG_ACTION='-1:0:PREPARE:FAIL'");
|
|
|
|
# Force a query plan that will have three fragments or more.
|
|
try:
|
|
self.client.execute("SELECT COUNT(*) FROM functional.alltypes a JOIN [SHUFFLE] \
|
|
functional.alltypes b on a.id = b.id")
|
|
assert "Query should have thrown an error"
|
|
except ImpalaBeeswaxException:
|
|
pass
|
|
|
|
for v in verifiers:
|
|
# Long timeout required because fragments may be blocked while sending data. The
|
|
# default value of --datastream_sender_timeout_ms is 120s before they wake up and
|
|
# cancel themselves.
|
|
#
|
|
# TODO: Fix when we have cancellable RPCs.
|
|
v.wait_for_metric(self.IN_FLIGHT_FRAGMENTS, 0, timeout=125)
|