1
0
mirror of synced 2025-12-23 03:47:05 -05:00
Commit Graph

2421 Commits

Author SHA1 Message Date
Airbyte
6ee9b1018c 🐙 destination-astra: run up-to-date pipeline [2024-09-21] (#45829) 2024-09-22 03:18:16 +03:00
Airbyte
c0a08c31d1 🐙 destination-qdrant: run up-to-date pipeline [2024-09-21] (#45830) 2024-09-22 03:18:13 +03:00
Airbyte
560111b38c 🐙 destination-cumulio: run up-to-date pipeline [2024-09-21] (#45831) 2024-09-22 03:18:10 +03:00
Airbyte
5f730ab273 🐙 destination-xata: run up-to-date pipeline [2024-09-21] (#45837) 2024-09-22 03:17:53 +03:00
Stephane Geneix
95430fba76 destination-dev-null: add to OSS registry (#45715) 2024-09-20 16:54:34 -07:00
Gireesh Sreepathi
befe6789b6 Destination Iceberg: initial release for cloud registry generation (#45711) 2024-09-21 02:39:46 +03:00
Stephane Geneix
d4952010c5 destination-dev-null: merge cloud and OSS (#45651) 2024-09-20 15:38:31 -07:00
Gireesh Sreepathi
965b8617e4 Destination Iceberg: Add support for AWS Glue Catalog (#45707) 2024-09-20 15:21:11 -07:00
Stephane Geneix
80faccf62b remove destination-e2e-test (#45702) 2024-09-20 15:18:09 -07:00
Stephane Geneix
0fc5bf6881 destination-dev-null: upgrade CDK (#45650) 2024-09-20 14:57:17 -07:00
Stephane Geneix
c6ec55deb1 destination-dev-null: convert test code to kotlin (#45649)
## What
<!--
* Describe what the change is solving. Link all GitHub issues related to this change.
-->

## How
<!--
* Describe how code changes achieve the solution.
-->

## Review guide
<!--
1. `x.py`
2. `y.py`
-->

## User Impact
<!--
* What is the end result perceived by the user?
* If there are negative side effects, please list them. 
-->

## Can this PR be safely reverted and rolled back?
<!--
* If unsure, leave it blank.
-->
- [ ] YES 💚
- [ ] NO 
2024-09-20 14:38:52 -07:00
Edward Gao
adb7e97800 Destination s3: Handle nonstandard fields in jsonschema (#45700) 2024-09-20 21:51:11 +03:00
Stephane Geneix
2e61b5a946 destination-dev-null: convert production code to kotlin (#45648)
## What
<!--
* Describe what the change is solving. Link all GitHub issues related to this change.
-->

## How
<!--
* Describe how code changes achieve the solution.
-->

## Review guide
<!--
1. `x.py`
2. `y.py`
-->

## User Impact
<!--
* What is the end result perceived by the user?
* If there are negative side effects, please list them. 
-->

## Can this PR be safely reverted and rolled back?
<!--
* If unsure, leave it blank.
-->
- [ ] YES 💚
- [ ] NO 
2024-09-20 11:06:29 -07:00
Stephane Geneix
bb3c27b817 destination-snowflake: update all dependencies (#45437) 2024-09-19 08:46:37 -07:00
Stephane Geneix
ccd0ea6c42 destination-snowflake: truncate large records (#45431)
## What
This change implements a large record truncation mechanism for the Snowflake destination connector to handle records exceeding Snowflake's 16MB row size limit.

## How
- The truncator preserves primary key fields and truncates other fields to fit within the 16MB limit.
- Added metadata to indicate which fields were truncated due to size limitations.


## User Impact
Users can now sync large records to Snowflake without encountering errors due to row size limitations. Fields may be truncated to fit within the 16MB limit, but primary keys are always preserved. Metadata is added to indicate which fields were affected.
2024-09-19 08:17:29 -07:00
Stephane Geneix
fcbbd43e67 destination-s3: fix issue with columnless streams (#45402)
Added support for handling empty data fields in Avro records and introduced a new test case for streams with no columns.

We've agreed there's a bug in a source (uscensus) that returns a single columnless stream. There's also a bug in the platform that allows a customer to create a connection using that columnless empty stream.
While it would be ideal for the destination connector to send an (yet non existent) `UpstreamError` back to the platform, we have some configurations of S3 (CSV, Json) that just allow teh columnless records to be persisted. We're just bringing the parquet/avro to the same permissiveness
2024-09-19 08:17:13 -07:00
Stephane Geneix
14f9e5c32d destination-snowflake: speed up metadata queries (#45422) 2024-09-18 15:14:54 -07:00
Stephane Geneix
0c9805e4e9 destination-s3: update all dependencies (#45436) 2024-09-18 14:52:40 -07:00
Stephane Geneix
21ec465d47 destination-postgres: update all dependencies (#45434)
## What
<!--
* Describe what the change is solving. Link all GitHub issues related to this change.
-->

## How
<!--
* Describe how code changes achieve the solution.
-->

## Review guide
<!--
1. `x.py`
2. `y.py`
-->

## User Impact
<!--
* What is the end result perceived by the user?
* If there are negative side effects, please list them. 
-->

## Can this PR be safely reverted and rolled back?
<!--
* If unsure, leave it blank.
-->
- [ ] YES 💚
- [ ] NO 
2024-09-18 14:41:26 -07:00
Stephane Geneix
a1bcb79852 destination-redshift: update all dependencies (#45435) 2024-09-18 14:20:41 -07:00
Stephane Geneix
d94b348c17 destination-databricks: update all dependencies (#45438) 2024-09-18 14:19:22 -07:00
Stephane Geneix
fea5ca8414 destination-bigquery: update all dependencies (#45433) 2024-09-18 14:14:41 -07:00
Stephane Geneix
7c8d612c51 destination-snowflake: fix destination state table collation (#38585)
when creating a table in snowflake, we use the DEFAULT_COLLATION set at the database or schema level. We should always use the utf-8 collation (which is the snowflake default), so our queries are simpler and faster (and accept more than 50 constants in an IN clause)
2024-09-18 11:46:40 -07:00
Stephane Geneix
e931c2add7 destination-snowflake: get tests to pass - durably (#45370)
### TL;DR

Make destination-snowflake pass all tests

### What changed?

- Updated CDK version to 0.45.0
- Reduced JUnit method execution timeout to 20 minutes
- Improved error handling in SnowflakeDestination's main function
- Enhanced error message for invalid permissions in integration test
- Implemented a more robust cleanup process for Airbyte internal tables and schemas
- Removed unused Batch and LocalFileBatch classes
- Not in the PR: I also deleted about 5k tables and 2k schemas, which were making our tests run slower than necessary. The cleanup logic will automate those cleanups. 

### How to test?

1. Run integration tests for the Snowflake destination connector
2. Verify that the new error message is displayed when testing with invalid permissions
3. Check that the cleanup process removes old tables and schemas as expected
4. Ensure that all existing functionality remains intact

### Why make this change?

These changes aim to improve the reliability and maintainability of the Snowflake destination connector. The updated CDK version and reduced test timeout should lead to faster and more efficient testing. The enhanced error handling and cleanup processes will help in identifying issues more quickly and keeping the test environment clean. Removing unused classes reduces code clutter and improves overall code quality.
2024-09-18 11:19:13 -07:00
Gireesh Sreepathi
daee3b69aa Destination Iceberg: Fix tests to run in airbyte-ci (#45206) 2024-09-17 07:53:05 -07:00
Aldo Gonzalez
9ae2cbee60 feat(Destination PGVector): new connector (#45428)
Co-authored-by: Aaron Steers <aj@airbyte.io>
2024-09-16 10:34:30 -06:00
Airbyte
3d84c481a0 🐙 destination-rabbitmq: run up-to-date pipeline [2024-09-14] (#45293) 2024-09-15 03:23:51 +03:00
Airbyte
f741b19587 🐙 destination-duckdb: run up-to-date pipeline [2024-09-14] (#45480) 2024-09-15 03:23:12 +03:00
Airbyte
2a97f35bd3 🐙 destination-vectara: run up-to-date pipeline [2024-09-14] (#45481) 2024-09-15 03:23:09 +03:00
Airbyte
98aa7dcfbb 🐙 destination-snowflake-cortex: run up-to-date pipeline [2024-09-14] (#45489) 2024-09-15 03:22:39 +03:00
Airbyte
0099c63059 🐙 destination-pinecone: run up-to-date pipeline [2024-09-14] (#45490) 2024-09-15 03:22:35 +03:00
Airbyte
1f721c176c 🐙 destination-typesense: run up-to-date pipeline [2024-09-14] (#45491) 2024-09-15 03:22:32 +03:00
Airbyte
3699c5bb79 🐙 destination-astra: run up-to-date pipeline [2024-09-14] (#45498) 2024-09-15 03:22:10 +03:00
Airbyte
d29ebda9e2 🐙 destination-xata: run up-to-date pipeline [2024-09-14] (#45516) 2024-09-15 03:21:14 +03:00
Airbyte
0b76f31358 🐙 destination-qdrant: run up-to-date pipeline [2024-09-14] (#45526) 2024-09-15 03:20:39 +03:00
Airbyte
0c65ad4bec 🐙 destination-databend: run up-to-date pipeline [2024-09-14] (#45531) 2024-09-15 03:20:22 +03:00
Airbyte
a1e280847c 🐙 destination-aws-datalake: run up-to-date pipeline [2024-09-14] (#45533) 2024-09-15 03:20:15 +03:00
Airbyte
37359fb3e0 🐙 destination-weaviate: run up-to-date pipeline [2024-09-14] (#45560) 2024-09-15 03:18:49 +03:00
Airbyte
e5cd6dea51 🐙 destination-firebolt: run up-to-date pipeline [2024-09-14] (#45562) 2024-09-15 03:18:43 +03:00
Airbyte
1e9ff52990 🐙 destination-timeplus: run up-to-date pipeline [2024-09-14] (#45568) 2024-09-15 03:18:24 +03:00
Airbyte
490d24e4f6 🐙 destination-milvus: run up-to-date pipeline [2024-09-14] (#45573) 2024-09-15 03:18:08 +03:00
Airbyte
7d5fbed77e 🐙 destination-google-sheets: run up-to-date pipeline [2024-09-14] (#45577) 2024-09-15 03:17:55 +03:00
Airbyte
38dec03b12 🐙 destination-cumulio: run up-to-date pipeline [2024-09-14] (#45580) 2024-09-15 03:17:46 +03:00
Edward Gao
8d04d840ba Destination databricks: move to intengrations (#45439) 2024-09-12 15:51:06 -07:00
Gireesh Sreepathi
bb1833cbf5 Destination Databricks: Create namespace if missing in CHECK (#45208) 2024-09-09 11:02:47 -07:00
Jove Zhong
06f8a2bca9 Destination Timeplus: fix issues in README and upgrade timeplus and cdk (#44758)
Co-authored-by: Marcos Marx <marcosmarxm@users.noreply.github.com>
2024-09-08 14:17:52 -03:00
Airbyte
4db089b000 🐙 destination-xata: run up-to-date pipeline [2024-09-07] (#45213) 2024-09-08 03:22:35 +03:00
Airbyte
548908819b 🐙 destination-weaviate: run up-to-date pipeline [2024-09-07] (#45216) 2024-09-08 03:22:26 +03:00
Airbyte
647bca3413 🐙 destination-qdrant: run up-to-date pipeline [2024-09-07] (#45217) 2024-09-08 03:22:24 +03:00
Airbyte
9f85142e6a 🐙 destination-timeplus: run up-to-date pipeline [2024-09-07] (#45218) 2024-09-08 03:22:20 +03:00