* Bump destination-postgres to 0.3.21
Goal is to trigger a publish of the arm64 build
* Update destination/postgres changelog
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* Added socket_timeout property for destination-clickhouse, jdbc default of 30 sec is to low
* correct unit test
* Added socket_timeout property for destination-clickhouse, jdbc default of 30 sec is to low
* correct unit test
* bump connector version and udpate doc
* auto-bump connector version
Co-authored-by: marcosmarxm <marcosmarxm@gmail.com>
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* BigQuery Denormalized "airbyte_type": "big_integer" to INT64
* updated changelog
* added unit test
* removed star import
* fixed checkstyle
* bump version
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* Remove "additionalProperties": false from spec for connectors with staging
* Remove "additionalProperties": false from spec for Redshift destination
* bump versions
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* start
* revert
* azblob
* bq
* bq denorm
* megapublish baaaabyyyy
* fix needs
* matrix connectors
* auto-bump connector version
* dont failfast and max parallel 5
* multi runno
* minor
* testing matrix agents
* name
* testing multi agents
* tmp fix
* new multi agents
* multi test
* tryy
* let's do this
* magico
* fix
* label test
* couple more connector bumps
* temp
* things
* check this
* lets gooo
* more connectors
* Delete TEMP-testing-command.yml
* auto-bump connector version
* added comment describing bash part
* running single thread
* catch sentry cli
* auto-bump connector version
* destinations
* + snowflake
* saved
* auto-bump connector version
* auto-bump connector version
* java source bumps
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* remove twice-defined methods
* label things
* revert action
* using the new test action
* point at action
* wrong tag on action
* update pool label
* update to use new ec2-github-runner fork
* this needs to be more generic than publisher
* change publish to run on pool
* add comment about runner-pool usage
* updated publish command docs for multi & parallel connector runs
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* unbump failed publish versions
* missed dockerfiles
* remove failed docs
* mssql fix
* overhauled the git comment output
* bumping a test connector that should work
* slight order switcheroo
* output connectors properly in first message
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* Removed part_size from connectors that use StreamTransferManager
* fixed S3DestinationConfigTest
* fixed S3JsonlFormatConfigTest
* upadate changelog and bump version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* auto-bump connector version
* upadate changelog and bump version for Redshift and Snowflake destinations
* auto-bump connector version
* fix GCS staging test
* fix GCS staging test
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* S3 destination: Updating processing data types for Avro/Parquet formats
* S3 destination: handle comparing data types
* S3 destination: clean code
* S3 destination: clean code
* S3 destination: handle case with unexpected json schema type
* S3 destination: clean code
* S3 destination: Extract the same logic for Avro/Parquet formats to separate parent class
* S3 destination: clean code
* S3 destination: clean code
* GCS destination: Update data types processing for Avro/Parquet formats
* GCS destination: clean redundant code
* S3 destination: handle case with numbers inside array
* S3 destination: clean code
* S3 destination: add unit test
* S3 destination: update unit test cases with number types.
* S3 destination: update unit tests.
* S3 destination: bump version for s3 and gcs
* auto-bump connector version
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* initial changes
* Edited google Sheets doc
* More edits
* edited the intro and prereqs for BigQuery
* edited the data loading section
* more edits
* Grammatical edits
* Formatting edits
* Rename databricks connector
* Rename connector in the seed
* Update changelog with pr id
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* Run gcs destination check all the time
* Add missing permissions to the status message
* Add more error messages
* Bump bigquery version
* auto-bump connector version
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* fix#12680: use correct max size for SUPER type
* address review comments
* docs: add changelog entry
* bump version
* test: make large record actually too big
* fix: use a method instead of member variable for correct override
* fix: if overall object is invalid, don't check individual elements
* feat: add unit test for isValidData
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>
* airbyte: Allow configuring PLAIN SASL mechanism for Kafka dest connector through WebApp
* bump up the version of kafka dest connector
* bump up the image tag for kafka destination connector in defs
* update _specs for kafka destination connector to match defs; SeedConnectorSpecs did not update it automatically
* Migrate to public databricks jdbc driver
* Update documentation
* Bump version
* Format code
* Check in databricks in seed file
* Check in databricks 0.2.0
* auto-bump connector version
Co-authored-by: Octavia Squidington III <octavia-squidington-iii@users.noreply.github.com>