Data Ingestion into BigQuery from Cloud Storage (Solution)

Data Ingestion into BigQuery from Cloud Storage (Solution)

Overview

  • Labs are timed and cannot be paused. The timer starts when you click Start Lab.

  • The included cloud terminal is preconfigured with the gcloud SDK.

  • Use the terminal to execute commands and then click Check my progress to verify your work.

Challenge scenario

You are managing Google BigQuery, a data warehouse service that lets you store, manage, and analyze large datasets. In this scenario, you need to create a dataset and a table within BigQuery to store employee details. The dataset will act as a container for your tables, while the table will hold the actual employee information.

  • You need to complete the following tasks:

    • Create a big query dataset: work_day

    • Create a table with employee the following schema details:

      column Type
      employee_id INTEGER
      device_id STRING
      username STRING
      department STRING
      office STRING
    • Import the csv data in your newly created table from pre-created cloud storage bucket named as qwiklabs-gcp-02-a85ba8626654-a1f8-bucket. The precreated bucket already has employees.csv file.

Click Check my progress to verify the objective.

Create BigQuery Schema and upload csv data


Solution of Lab

https://www.youtube.com/watch?v=n3matiOxKwc

```apache curl -LO raw.githubusercontent.com/ePlus-DEV/storage/refs/heads/main/labs/data-ingestion-into-bigquery-from-cloud-storage-solution/lab.sh source lab.sh ```

Script Alternative

export BUCKET=
bq mk work_day && bq load --source_format=CSV --skip_leading_rows=1 work_day.employee gs://$BUCKET/employees.csv employee_id:INTEGER,device_id:STRING,username:STRING,department:STRING,office:STRING