* new source: kafka * refactored util and config classes * fixed discover method for configs based on topic pattern
Kafka Source
This is the repository for the Kafka source connector. For information about how to use this connector within Airbyte, see the documentation.
Local development
Prerequisites
- If you are using Python for connector development, minimal required version
= 3.7.0 - Valid credentials (see the "Create credentials section for instructions) TODO: which languages and tools does a user need to develop on this connector? add them to the bullet list above
Iteration
TODO: which commands should a developer use to run this connector locally?
Testing
Unit Tests
TODO: how can a user run unit tests?
Integration Tests
TODO: how can a user run integration tests? this section is currently under construction -- please reach out to us on Slack for help with setting up Airbyte's standard test suite
Locally running the connector docker image
First, make sure you build the latest Docker image:
docker build . -t airbyte/kafka:dev
Then run any of the connector commands as follows:
docker run --rm airbyte/source-kafka:dev spec
docker run --rm -v $(pwd)/secrets:/secrets airbyte/source-kafka:dev check --config /secrets/config.json
docker run --rm -v $(pwd)/secrets:/secrets airbyte/source-kafka:dev discover --config /secrets/config.json
docker run --rm -v $(pwd)/secrets:/secrets -v $(pwd)/sample_files:/sample_files airbyte/source-kafka:dev read --config /secrets/config.json --catalog /sample_files/configured_catalog.json
Create credentials
If you are a community contributor, follow the instructions in the documentation
to generate the necessary credentials. Then create a file secrets/config.json conforming to the spec.json file. secrets is gitignored by default.
If you are an Airbyte core member, copy the credentials from Lastpass under the secret name source kafka test creds
and place them into secrets/config.json.