* monitor "schemas" queue to run refresh_schema
`refresh_schema` tasks won't run because "schemas" queue isn't consumed
with default settings.
and it cause leaking redis storage, a "schemas" list is growing with time.
this PR fix it, adds "schemas" queue to monitor by celery.
* use scheduled_worker for "schemas" queue
instead of "adhoc_worker"
* open port 3000 for remote debugging
* add ptvsd
* use port 5678 to avoid changes in VSCode's default config
* attach to ptvsd
* no need to wait for attach
* actually, --debugger seems to be working
* create a new docker entry point for remote debugging
* alternative method to switch to debugging
By default Celery will use a file celerybeat-schedule in the current directory.
This is an issue in a Kubernetes/Openshift environment as the file may be lost or even impossible to write.
Allow $REDASH_WEB_WORKERS to be set in the environment to change the
number of Gunicorn workers that are started (currently hardcoded to
four). If not set, the default is four, so this will not affect
existing users at all.
Documentated by example in docker-compose example manifests.
1. Simplify user creation to create a non system user (so the uid will usually
match the host user).
2. Set the user to redash & remove the need to change user in docker entrypoint.