Tailer Documentation
  • What is Tailer Platform?
  • Getting Started
    • Prepare your local environment for Tailer
    • Install Tailer SDK
    • Set up Google Cloud Platform
    • Encrypt your credentials
  • [Tutorial] Create a first data pipeline
    • Introduction
    • Prepare the demonstration environment
    • Copy files from one bucket to another
    • Load files into BigQuery tables
    • Prepare data
    • Build predictions
    • Export data
    • Congratulations!
    • [Video] Automatic Script
      • SQL script file
      • DDL script file
      • Tables to Tables script file
      • Launch configuration and furthermore
  • Data Pipeline Operations
    • Overview
    • Set constants with Context
      • Context configuration file
    • Move files with Storage to Storage
      • Storage to Storage configuration file
    • Load data with Storage to Tables
      • Storage to Tables configuration file
      • Storage to Tables DDL files
    • Stream incoming data with API To Storage
      • API To Storage configuration file
      • API To Storage usage examples
    • Transform data with Tables to Tables
      • Tables to Tables configuration file
      • Table to Table SQL and DDL files
    • Export data with Tables to Storage
      • [V3] Table to Storage configuration file
      • Table to Storage SQL file
      • [V1-V2: deprecated] Table to Storage configuration file
    • Orchestrate processings with Workflow
      • [V2] Workflow configuration file
      • [V1: deprecated] Workflow configuration file
    • Convert XML to CSV
      • Convert XML to CSV configuration file
    • Use advanced features with VM Launcher
      • Process code with VM Launcher
        • VM Launcher configuration file for code processing
      • Encrypt/Decrypt data with VM Launcher
        • VM Launcher configuration file for data encryption
        • VM Launcher configuration file for data decryption
    • Monitoring and Alerting
      • Monitoring and alerting parameters
    • Asserting Data quality with Expectations
      • List of Expectations
    • Modify files with File Utilities
      • Encrypt/Decrypt data with File Utilities
        • Configuration file for data encryption
        • Configuration file for data decryption
    • Transfer data with GBQ to Firestore
      • Table to Storage: configuration file
      • Table to Storage: SQL file
      • VM Launcher: configuration file
      • File-to-firestore python file
  • Tailer Studio
    • Overview
    • Check data operations' details
    • Monitor data operations' status
    • Execute data operations
    • Reset Workflow data operations
    • Archive data operations
    • Add notes to data operations and runs
    • View your data catalog
    • Time your data with freshness
  • Tailer API
    • Overview
    • Getting started
    • API features
  • Release Notes
    • Tailer SDK Stable Releases
    • Tailer Beta Releases
      • Beta features
      • Beta configuration
      • Tailer SDK API
    • Tailer Status
Powered by GitBook
On this page
  • What is a GBQ to Firestore configuration?
  • ✅ Overall operations needed
  • ⚙️ How it works
  • 📋 How to deploy a GBQ to Firestore data pipeline:

Was this helpful?

Edit on GitHub
  1. Data Pipeline Operations

Transfer data with GBQ to Firestore

Learn how to transfer a Google BigQuery table contents into Firestore documents.

PreviousConfiguration file for data decryptionNextTable to Storage: configuration file

Last updated 1 year ago

Was this helpful?

What is a GBQ to Firestore configuration?

A GBQ to Firestore data pipeline allows you to transform each row of a BigQuery table into Firestore documents.

✅ Overall operations needed

GBQ to Firestore is an advanced data pipeline which requires prior knowledge of:

  • Python

⚙️ How it works

  1. A tables-to-storage data operation is scheduled, or triggered by an event (for example a tables-to-tables successful run)

  2. The SQL query you specified in your tables-to-storage data operation is executed to extract the relevant data from BigQuery into a JSON file located in the Google Cloud Storage bucket of your choice

  3. Then a vm-launcher data operation is triggered to launch a VM on Google Compute Engine to execute the Python script that loads the JSON file into Firestore

  4. Once the execution is complete, the VM is stopped automatically

  5. Lastly, you can check the file tree in Firestore and see your data!

📋 How to deploy a GBQ to Firestore data pipeline:

You must follow these steps:

The following pages describe how to deploy a first end-to-end GBQ to Firestore data pipeline.

💡
⚠️
Tables to storage configuration
VM-launcher configuration
Workflow configuration
(Data structures - Dictionaries)
Firestore
Deploy a tables to storage data operation
Deploy a vm-launcher data operation for code processing
Deploy a workflow data operation