Export data with Tables to Storage

Learn how to export data located in a BigQuery table into CSV/JSON files using a Table to Storage operation.

​💡 What is Table to Storage?

A Table to Storage (TTS) data pipeline operation allows you to export your data from a BigQuery table to a CSV/JSON file in a Google Cloud Storage bucket so you can leverage them with other tools, such as a warehouse management system.

âś… Supported file types


  • Google BigQuery

Export files

  • CSV file in Google Cloud Storage

  • JSON file in Google Cloud Storage

⚙️ How it works

When the Table to Storage Tailer workflow is triggered by an event (usually a BigQuery table update):

  • The SQL query you specify will be executed to extract the relevant data from the source BigQuery table.

  • The data will be exported to a CSV file or a JSON file located in the GCS bucket you specified.

đź“‹ How to deploy a Table to Storage data operation

  1. Access your tailer folder (created during installation).

  2. Create a working folder as you want, and create a JSON file for your data operation inside.

  3. Create a SQL file to determine what data to extract.

  4. Prepare your JSON configuration file. Refer to this page to learn about all the parameters.

  5. Create a Workflow configuration file that will define how to trigger it.

  6. Access your working folder by running the following command:

    cd "[path to your working folder]"
  7. To deploy the data operation, run the following command:

    tailer deploy configuration your-configuration.json
  8. Log in to Tailer Studio to check the status and details of your data operation.

  9. For your workflow to be executed, you either need to run the data operation that is set to trigger it in your Workflow data operation (previous step in your data pipeline), or to launch it manually from Tailer Studio.

  10. Access the GCS bucket to check your output file (CSV or JSON).