Ends in
00
days
00
hrs
00
mins
00
secs
ENROLL NOW

đŸŒ¸ 25% OFF All Reviewers on our International Women's Month Sale Extension! Save 10% OFF All Subscription Plans

Google BigQuery

Home » Google Cloud » Google BigQuery

Google BigQuery

Last updated on March 13, 2026

Google BigQuery Cheat Sheet 

  • A fully managed data warehouse where you can feed petabyte-scale data sets and run SQL-like queries.

Features

  • Cloud BigQuery is a serverless data warehousing technology.
  • It provides integration with the Apache big data ecosystem allowing Hadoop/Spark and Beam workloads to read or write data directly from BigQuery using Storage API.
  • BigQuery supports a standard SQL dialect that is ANSI:2011 compliant, which reduces the need for code rewrites.
  • Automatically replicates data and keeps a seven-day history of changes which facilitates restoration and data comparison from different times.
  • Gemini in BigQuery AI assistance is now included in BigQuery pricing models.
  • Apache Iceberg tables support via BigLake for streaming, advanced analytics, and AI use cases.
  • BigQuery AI: Train, evaluate, and run ML models (linear regression, k-means, time series) directly within BigQuery using SQL. Integrate with Vertex AI Model Registry for MLOps.
  • AI agents: Data Engineering Agent (automate data prep), Data Science Agent (streamline ML lifecycle), Conversational Analytics Agent (ask questions in plain language).
  • Tutorials dojo strip
  • Dataplex Universal Catalog for governance: automatic metadata harvesting, data profiling, data quality, and lineage.
  • Cross-region disaster recovery with dataset replication for protection against region outages.

Loading data into BigQuery

You must first load your data into BigQuery before you can run queries. To do this you can:

  • Load a set of data records from Cloud Storage or from a local file. The records can be in Avro, CSV, JSON (newline delimited only), ORC, or Parquet format.
  • Export data from Datastore or Firestore and load the exported data into BigQuery.
  • Load data from other Google services, such as
    • Google Ad Manager
    • Google Ads
    • Google Play
    • Cloud Storage
    • Youtube Channel Reports
    • Youtube Content Owner reports
  • Stream data one record at a time using streaming inserts.
  • Write data from a Dataflow pipeline to BigQuery.
  • Use DML statements to perform bulk inserts. Note that BigQuery charges for DML queries. See Data Manipulation Language pricing.

Querying from external data sources

  • BigQuery offers support for querying data directly from:
    • Cloud BigTable
    • Cloud Storage
    • Cloud SQL
  • Supported formats are:
    • Avro
    • CSV
    • JSON (newline delimited only)
    • ORC
    • Parquet
  • To query data on external sources, you have to create external table definition file that contains the schema definition and metadata.

Google BigQuery Monitoring

  • BigQuery creates log entries for actions such as creating or deleting a table, purchasing slots, or running a load job.

Google BigQuery Pricing

  • Free tier (monthly):
    • Storage: Up to a certain amount free
    • Queries: Up to a certain amount of on-demand compute free
  • Compute (analysis):
    • On-demand: Billed per TiB scanned (first amount free). Access to a set number of concurrent slots shared per project.
    • Editions (Standard, Enterprise, Enterprise Plus): Billed per slot hour. Includes Gemini in BigQuery AI assistance features.
  • Storage:
    • Logical storage: Billed per GiB (uncompressed bytes, tables modified in recent period). First amount free.
    • Physical storage: Billed per GiB (compressed bytes, tables modified for extended period). First amount free.
  • Data ingestion:
    • Batch loading from Cloud Storage: Free (using shared slot pool)
    • Streaming inserts: Billed per amount ingested (minimum row size applies)
    • Storage Write API: Billed per GiB (first amount free per month)
  • Data extraction:
    • Batch export to Cloud Storage: Free (using shared slot pool)
    • Streaming reads (Storage Read API): Billed per TiB read

Loading and exporting data are free. All charges are prorated per second.

For current pricing details, refer to the official Google Cloud BigQuery pricing page.

Validate Your Knowledge

Question 1

Your company has a 5 TB file in Parquet format stored in Google Cloud Storage bucket. A team of analysts, who are only proficient in SQL, needs to temporarily access these files to run ad-hoc queries. You need a cost-effective solution to fulfill their request as soon as possible.

What should you do?

  1. Load the data in a new BigQuery table. Use the bq load command, specify PARQUET using the --source_format flag, and include a Cloud Storage URL.
  2. Create external tables in BigQuery. Use the Cloud Storage URL as a data source.
  3. Load the data in BigTable. Give the analysts the necessary IAM roles to run SQL queries.
  4. Import the data to Memorystore to provide quick access to Parquet data in the Cloud Storage bucket.

Correct Answer: 2

An external data source (also known as a federated data source) is a data source that you can query directly even though the data is not stored in BigQuery. Instead of loading or streaming the data, you create a table that references the external data source.

BigQuery supports querying Cloud Storage data in the following formats:

– Comma-separated values (CSV)

– JSON (newline-delimited)

– Avro

– ORC

Free AWS Courses

– Parquet

– Datastore exports

– Firestore exports

BigQuery supports querying Cloud Storage data from these storage classes:

– Standard

– Nearline

– Coldline

– Archive

To query a Cloud Storage external data source, provide the Cloud Storage URL path to your data, and create a table that references the data source. The table used to reference the Cloud Storage data source can be a permanent table or a temporary table.

It is stated in the scenario that a low-cost and temporary access to Parquet data should be provided. Using the BigQuery temporary external table will satisfy this requirement compared to loading the data to permanent tables that use datasets to store the data. Querying an external data source using a temporary table is useful for one-time, ad-hoc queries over external data, or for extract, transform, and load (ETL) processes.

Hence, the correct answer is: Create external tables in BigQuery. Use the Cloud Storage URL as a data source.

The option that says: Load the data in a new BigQuery table. Use the bq load command, specify PARQUET using the –source_format flag, and include a Cloud Storage URL is incorrect because doing this will load the data on the BigQuery dataset which is not ideal for accessing data temporarily. Instead, you can use the temporary table for external data sources in BigQuery.

The option that says: Load the data in BigTable. Give the analysts the necessary IAM roles to run SQL queries is incorrect because BigTable is a NoSQL database. Note: it is stated in the scenario that the analysts are only proficient in SQL, and BigTable is not a type of SQL database.

The option that says: Import the data to Memorystore to provide quick access to Parquet data in the Cloud Storage bucket is incorrect because Memorystore is only used to build application caches. This service is compatible with open source Redis and Memcached.

References:

https://cloud.google.com/bigquery/external-data-cloud-storage
https://cloud.google.com/bigquery/external-data-sources

Note: This question was extracted from our Google Certified Associate Cloud Engineer Practice Exams.

For more Google Cloud practice exam questions with detailed explanations, check out the Tutorials Dojo Portal:

Google Certified Associate Cloud Engineer Practice Exams

Google BigQuery Cheat Sheet References:

https://cloud.google.com/bigquery
https://cloud.google.com/bigquery/docs/introduction

đŸŒ¸ 25% OFF All Reviewers on our International Women’s Month Sale Extension! Save 10% OFF All Subscription Plans & 5% OFF Store Credits/Gift Cards!

Tutorials Dojo portal

Learn AWS with our PlayCloud Hands-On Labs

$2.99 AWS and Azure Exam Study Guide eBooks

tutorials dojo study guide eBook

New AWS Generative AI Developer Professional Course AIP-C01

AIP-C01 Exam Guide AIP-C01 examtopics AWS Certified Generative AI Developer Professional Exam Domains AIP-C01

Learn GCP By Doing! Try Our GCP PlayCloud

Learn Azure with our Azure PlayCloud

FREE AI and AWS Digital Courses

FREE AWS, Azure, GCP Practice Test Samplers

SAA-C03 Exam Guide SAA-C03 examtopics AWS Certified Solutions Architect Associate

Subscribe to our YouTube Channel

Tutorials Dojo YouTube Channel

Follow Us On Linkedin

Written by: Jon Bonso

Jon Bonso is the co-founder of Tutorials Dojo, an EdTech startup and an AWS Digital Training Partner that provides high-quality educational materials in the cloud computing space. He graduated from MapĂºa Institute of Technology in 2007 with a bachelor's degree in Information Technology. Jon holds 10 AWS Certifications and is also an active AWS Community Builder since 2020.

AWS, Azure, and GCP Certifications are consistently among the top-paying IT certifications in the world, considering that most companies have now shifted to the cloud. Earn over $150,000 per year with an AWS, Azure, or GCP certification!

Follow us on LinkedIn, YouTube, Facebook, or join our Slack study group. More importantly, answer as many practice exams as you can to help increase your chances of passing your certification exams on your first try!

View Our AWS, Azure, and GCP Exam Reviewers Check out our FREE courses

Our Community

~98%
passing rate
Around 95-98% of our students pass the AWS Certification exams after training with our courses.
200k+
students
Over 200k enrollees choose Tutorials Dojo in preparing for their AWS Certification exams.
~4.8
ratings
Our courses are highly rated by our enrollees from all over the world.

What our students say about us?