Monday, June 5, 2023
Search Console bulk data export is a powerful way to get your website's search performance data into BigQuery to increase storage, analysis, and reporting capabilities. For example, after exporting the data, you can perform query and URL clustering, run analyses on long-tail search queries, and join search with other sources of data. You can also choose to retain the data for as long as you need it.
When using Bulk data exports, it's important to make informed decisions when managing the data processing and storage costs. There are no costs associated with Search Console to export the data; however, do read the BigQuery pricing to understand what you will be billed for. In this post, we'll discuss tips to help you take advantage of the new data without incurring significant cost.
If you haven't set up a bulk data export yet, check the step-by-step guide in the Search Console help center. For an overview of the data available through the export, check the video embedded here.
Create billing alerts and restrictions
When considering your costs, it might help to think through how much you'd be willing to spend. The answer to that question is likely to be different between storage, analysis, and monitoring. For example, you might be willing to pay a certain amount to make sure you're storing all your data, but less to create a reporting platform. While thinking through that, you might want to set a monthly budget to invest in Search data.
Once you have a budget amount in mind, you can create a Google Cloud budget alert to avoid surprises on your bill. You can also set threshold rules that trigger email notifications when you're advancing towards your budget amount.
For added protection, you can also restrict the number of bytes billed for a query. If you do that, the number of bytes that the query will read is estimated before the query execution. If the number of estimated bytes is beyond the limit, then the query fails without incurring a charge.
Don't build dashboards directly on raw data
BigQuery is fast, and it is tempting to link your dashboard directly to the Search Console exported tables. But for large sites, this dataset is very large (especially with over-time queries). If you build a dashboard that recomputes summary information on every view and share that within your company, this will quickly run up large query costs.
To avoid these costs, consider pre-aggregating the data from every daily drop and materializing one or more summary tables. Your dashboard can then query a much smaller time series table, decreasing processing costs.
Optimize data storage costs
When you start a bulk data export, by default, data is kept forever in your BigQuery dataset. However, you can update the default partition expiration times so that date partitions are automatically deleted after a year, or 16 months, or any duration you desire.
The data exported can be valuable to you, but it can be very large. Use your business knowledge and consider retaining it long enough for deep analyses, but not too long that it becomes a burden. One option is to keep a sampled version of older tables while keeping the entire table of more recent dates.
Optimize your SQL queries
While querying your Search Console data, you should make sure your queries are optimized for performance. If you're new to BigQuery, check the guidelines and sample queries in the help center. There are three techniques you should try.
1. Limit the input scan
First of all, avoid using
this is the most expensive way to query the data, BigQuery does a full scan of every column in the table.
LIMIT clause does not affect the amount of data read.