Link Search Menu Expand Document


Airbyte is an open-source platform for syncing data from applications, APIs, and databases to warehouses, lakes, and other destinations. You can use Airbyte’s connectors to get your data pipelines to consolidate many input sources.

The integration between Airbyte and lakeFS brings resilience and manageability when you use Airbyte connectors to sync data to your S3 buckets by leveraging lakeFS branches and atomic commits and merges.

Use cases

You can take advantage of lakeFS consistency guarantees and Data Lifecycle Management when ingesting data to S3 using lakeFS:

  1. Consolidate many data sources to a single branch and expose them to consumers simultaneously when merging to the main branch.
  2. Test incoming data for breaking schema changes using lakeFS hooks.
  3. Prevent consumers from reading partial data from connectors which failed half-way through sync.
  4. Experiment with ingested data on a branch before exposing it.

S3 Connector

lakeFS exposes an S3 Gateway that enables applications to communicate with lakeFS the same way they would with Amazon S3. You can use Airbyte’s S3 Connector to upload data to lakeFS.


If using Airbyte OSS, please ensure you are using S3 destination connector version 0.3.17 or higher. Previous connector versions are not supported.

Configuring lakeFS using the connector

Set the following parameters when creating a new Destination of type S3:

Name Value Example
Endpoint The lakeFS S3 gateway URL
S3 Bucket Name The lakeFS repository where the data will be written example-repo
S3 Bucket Path The branch and the path where the data will be written main/data/from/airbyte Where main is the branch name, and data/from/airbyte is the path under the branch.
S3 Bucket Region Not applicable to lakeFS, use us-east-1 us-east-1
S3 Key ID The lakeFS access key id used to authenticate to lakeFS. AKIAlakefs12345EXAMPLE
S3 Access Key The lakeFS secret access key used to authenticate to lakeFS. abc/lakefs/1234567bPxRfiCYEXAMPLEKEY

The UI configuration will look as follows:

S3 Destination Connector Configuration