mirror of
https://github.com/apache/impala.git
synced 2025-12-19 18:12:08 -05:00
sqlparse-0.1.19 is the last version of sqlparse that supports Python 2.6. Testing: - Ran all end-to-end tests Change-Id: Ide51ef3ac52d25a96b0fa832e29b6535197d23cb Reviewed-on: http://gerrit.cloudera.org:8080/10354 Reviewed-by: David Knupp <dknupp@cloudera.com> Tested-by: Impala Public Jenkins <impala-public-jenkins@cloudera.com>
81 lines
2.2 KiB
Python
81 lines
2.2 KiB
Python
# Copyright (C) 2008 Andi Albrecht, albrecht.andi@gmail.com
|
|
#
|
|
# This module is part of python-sqlparse and is released under
|
|
# the BSD License: http://www.opensource.org/licenses/bsd-license.php.
|
|
|
|
"""filter"""
|
|
|
|
from sqlparse import lexer
|
|
from sqlparse.engine import grouping
|
|
from sqlparse.engine.filter import StatementFilter
|
|
|
|
# XXX remove this when cleanup is complete
|
|
Filter = object
|
|
|
|
|
|
class FilterStack(object):
|
|
|
|
def __init__(self):
|
|
self.preprocess = []
|
|
self.stmtprocess = []
|
|
self.postprocess = []
|
|
self.split_statements = False
|
|
self._grouping = False
|
|
|
|
def _flatten(self, stream):
|
|
for token in stream:
|
|
if token.is_group():
|
|
for t in self._flatten(token.tokens):
|
|
yield t
|
|
else:
|
|
yield token
|
|
|
|
def enable_grouping(self):
|
|
self._grouping = True
|
|
|
|
def full_analyze(self):
|
|
self.enable_grouping()
|
|
|
|
def run(self, sql, encoding=None):
|
|
stream = lexer.tokenize(sql, encoding)
|
|
# Process token stream
|
|
if self.preprocess:
|
|
for filter_ in self.preprocess:
|
|
stream = filter_.process(self, stream)
|
|
|
|
if (self.stmtprocess or self.postprocess or self.split_statements
|
|
or self._grouping):
|
|
splitter = StatementFilter()
|
|
stream = splitter.process(self, stream)
|
|
|
|
if self._grouping:
|
|
|
|
def _group(stream):
|
|
for stmt in stream:
|
|
grouping.group(stmt)
|
|
yield stmt
|
|
stream = _group(stream)
|
|
|
|
if self.stmtprocess:
|
|
|
|
def _run1(stream):
|
|
ret = []
|
|
for stmt in stream:
|
|
for filter_ in self.stmtprocess:
|
|
filter_.process(self, stmt)
|
|
ret.append(stmt)
|
|
return ret
|
|
stream = _run1(stream)
|
|
|
|
if self.postprocess:
|
|
|
|
def _run2(stream):
|
|
for stmt in stream:
|
|
stmt.tokens = list(self._flatten(stmt.tokens))
|
|
for filter_ in self.postprocess:
|
|
stmt = filter_.process(self, stmt)
|
|
yield stmt
|
|
stream = _run2(stream)
|
|
|
|
return stream
|