1
0
mirror of synced 2026-01-04 09:04:47 -05:00
Files
airbyte/docusaurus/platform_versioned_docs/version-2.0/operator-guides/using-kestra-plugin.md
devin-ai-integration[bot] 34282926a1 docs: Generate version 2.0 of platform documentation (#68102)
## What

Generates version 2.0 of the Airbyte platform documentation using
Docusaurus's built-in versioning system. This creates a frozen snapshot
of the current documentation that users can reference.

Requested by ian.alton@airbyte.io via [Slack
thread](https://airbytehq-team.slack.com/archives/D08FX8EC9L0/p1760490197805979?thread_ts=1760490197.805979).

Link to Devin run:
https://app.devin.ai/sessions/689693593bac44f4903f476aa17b872e

## How

- Ran `pnpm run docusaurus docs:version:platform 2.0` in the docusaurus
directory
- This automatically:
- Created `platform_versioned_docs/version-2.0/` containing a snapshot
of all current platform docs
- Created `platform_versioned_sidebars/version-2.0-sidebars.json` with
the sidebar navigation structure
  - Updated `platform_versions.json` to add "2.0" to the version list
- Ran prettier to format the JSON files
- Verified the documentation builds successfully locally (build
completed in ~3 minutes with only pre-existing broken anchor warnings)

## Review guide

1. **Verify timing**: Confirm this is the correct time to release
version 2.0 of the documentation
2. **Version order**: Check `docusaurus/platform_versions.json` - verify
"2.0" is first in the array (newest version first)
3. **Build verification**: Ensure CI/Vercel builds pass without errors
4. **Spot check**: Optionally review 2-3 files in
`docusaurus/platform_versioned_docs/version-2.0/` to ensure content
looks reasonable

Note: This is a standard Docusaurus versioning operation that creates a
frozen snapshot of the current "next" documentation. The generated files
are extensive (500+ files) but follow Docusaurus conventions.

## User Impact

Users will see version 2.0 available in the version dropdown on
docs.airbyte.com. This provides a stable reference point for platform
documentation at this point in time. Existing versions (1.6, 1.7, 1.8)
remain unchanged.

## Can this PR be safely reverted and rolled back?

- [x] YES 💚

This is an additive change that doesn't modify existing versioned docs.
Reverting would simply remove version 2.0 from the version list and
delete the associated documentation files.

Co-authored-by: Devin AI <158243242+devin-ai-integration[bot]@users.noreply.github.com>
Co-authored-by: ian.alton@airbyte.io <ian.alton@airbyte.io>
2025-10-14 18:27:24 -07:00

5.2 KiB
Raw Blame History

description, products
description products
Using the Kestra Plugin to Orchestrate Airbyte oss-*

Using the Kestra Plugin

Kestra has an official plugin for Airbyte, including support for self-hosted Airbyte and Airbyte Cloud. This plugin allows you to trigger data replication jobs (Syncs) and wait for their completion before proceeding with any downstream tasks. Alternatively, you may also run those syncs in a fire-and-forget way by setting the wait argument to false.

After Airbyte tasks successfully ingest raw data, you can easily start running downstream data transformations with dbt, Python, SQL, Spark, and many more, using a variety of available plugins. Check the plugin documentation for a list of all supported integrations.

Available tasks

These are the two main tasks to orchestrate Airbyte syncs:

  1. The io.kestra.plugin.airbyte.connections.Sync task will sync connections for a self-hosted Airbyte instance

  2. The io.kestra.plugin.airbyte.cloud.jobs.Sync task will sync connections for Airbyte Cloud

1. Set up the tools

First, make sure you have Docker installed. We'll be using the docker-compose command, so your installation should contain docker-compose. When you use Docker Desktop, Docker Compose is already included.

Start Airbyte

If this is your first time using Airbyte, we suggest following the Quickstart Guide. When creating Airbyte connections intended to be orchestrated with Kestra, set your Connection's sync frequency to manual. Kestra will automate triggering Airbyte jobs in response to external events or based on a schedule youll provide.

Install Kestra

If you havent started Kestra yet, download the Docker Compose file:

curl -o docker-compose.yml https://raw.githubusercontent.com/kestra-io/kestra/develop/docker-compose.yml

Then, run docker compose up -d and navigate to the UI. You can start building your first flows using the integrated code editor in the UI.

airbyte_kestra_CLI

2. Create a flow from the UI

Kestra UI provides a wide range of Blueprints to help you get started.

Navigate to Blueprints. Then type "Airbyte" in the search bar to find the desired integration. This way, you can easily accomplish fairly standardized data orchestration tasks, such as the following:

  1. Run a single Airbyte sync on a schedule
  2. Run multiple Airbyte syncs in parallel
  3. Run multiple Airbyte syncs in parallel, then clone a Git repository with dbt code and trigger dbt CLI commands
  4. Run a single Airbyte Cloud sync on a schedule
  5. Run multiple Airbyte Cloud syncs in parallel, then clone a Git repository with dbt code and trigger dbt CLI commands
  6. Run multiple Airbyte Cloud syncs in parallel, then run a dbt Cloud job

Select a blueprint matching your use case and click "Use".

airbyte_kestra_blueprints

Then, within the editor, adjust the connection ID and task names and click "Save". Finally, trigger your flow.

3. Simple demo

Here is an example flow that triggers multiple Airbyte connections in parallel to sync data for multiple Pokémon.

id: airbyte_syncs
namespace: company.team
description: Gotta catch em all!

tasks:
  - id: data_ingestion
    type: io.kestra.core.tasks.flows.Parallel
    tasks:
      - id: charizard
        type: io.kestra.plugin.airbyte.connections.Sync
        connectionId: 9bb96539-73e7-4b9a-9937-6ce861b49cb9
      - id: pikachu
        type: io.kestra.plugin.airbyte.connections.Sync
        connectionId: 39c38950-b0b9-4fce-a303-06ced3dbfa75
      - id: psyduck
        type: io.kestra.plugin.airbyte.connections.Sync
        connectionId: 4de8ab1e-50ef-4df0-aa01-7f21491081f1

taskDefaults:
  - type: io.kestra.plugin.airbyte.connections.Sync
    values:
      url: http://host.docker.internal:8000/
      username: "{{ secret('AIRBYTE_USERNAME') }}"
      password: "{{ secret('AIRBYTE_PASSWORD') }}"

triggers:
  - id: every_minute
    type: io.kestra.core.models.triggers.types.Schedule
    cron: "*/1 * * * *"

Next steps

If you liked that demo, check out the blog post about using Airbyte and Kestra Terraform providers together to manage Everything as Code.

If you encounter anything unexpected while reproducing this tutorial, you can open a GitHub issue or ask via Kestra Community Slack. Lastly, give Kestra a GitHub star if you like the project.