What is the time taken by google analytics to start exporting historical data into the google cloud after the linking is complete between Big Query and Google Analytics.
According to Google's documentation:
Once the linkage is complete, data should
start flowing to your BigQuery project within 24 hours. 1 file will be
exported each day that contains the previous day’s data, and 3 files
will be exported each day that contain the current day's data. We will
provide a historical export of the smaller of 10 billion hits or 13
months of data within 4 weeks after the integration is complete.
Related
I have a problem with cost data uploaded to Google Analytics via the Management API.
The script itself works just fine. On a daily schedule it uploads a CSV from BigQuery to the management API. But in some cases the cost data shown in the UI is different from the data in the upload file.
For example:
In my BigQuery Table is see costs of 349.44 for a certain campaign on a certain date.
In the UI for the same campaign and date I see 92.31.
If I download the uploaded CSV again from the Data Upload section I see 257.13.
So three different numbers in three places that should not differ.
For the same campaign and other dates the data is correct.
All costs are in EUR so no currency conversion takes place and not just recent data is affected but also data from many month ago. I haven't found any resources that address this issue and would be greatful for any input.
I use Bigquery Sandbox and export raw event data from Google Analytics to Bigquery. But The number of events 2 places are different.Bigquery much less Google Analytics.
My number of events per day usually ranges from 1 to 200 thousand events meanwhile GA usually ranges 7-10 million evets day. And The storage space I have left is 6.97/10 GB.
I'm currently in the middle of the process of linking my Google analytics data to Big Query and the following note caught my attention when selecting the view to pick.
If this is the first time that you have linked this view, then data
will be backfilled for the smaller of 13 months or 10 billion hits.
Its a little unclear to me whether this 13 months of data will have costs in importing once I linked to BigQuery.
The process itself doesn't have a price component for back filling, but the storage it will occupy in BigQuery adds to your storage costs.
If you don't want old data, make sure you archive/remove/delete it.
We are having some issues pulling yesterday's Google Analytics data from BigQuery. Can anyone explain at what point a previous day's GA data is finalized?
There is some explanation here of the intraday tables, but it's not very clear:
https://support.google.com/analytics/answer/3437719?hl=en
To get previous day data do you need to need to use the intraday tables at all? Do you have access to the fully processed dataset at 8am local time? Or is it 8 hours after the current day UTC+14:00 (etc)?
I had a similar question and asked their support, this is the reply:
"According to this Google Analytics documentation , it states that '1 file will be exported each day that contains the previous day’s data, and 3 files will be exported each day that contain the current day's data'. In such, the minimum time that the data from Google Analytics to be exported to BigQuery was 8 hours. Although Google Analytics can be linked to BigQuery, the availability of data depends on how it was served by Google Analytics 360."
But based on experience, it's really a minimum time. Sometimes there are delays of 4-5 hours.
My team has been pressing Google's support for providing SLA's for BigQuery dump, so they updated the documentation:
This feature is not governed by a service-level agreement (SLA).
In practice we are experiencing regular delays anywhere between 2 to 12 hours.
I'm using the Google-Analytics API to query my analytics for data using the Google Analytics Spreadsheet Add-on We then use the spreadsheet data in Google Data studio for a dashboard to display the data.
Everything has been going well for the last few months, however over the last 48 hours we have begun to receive sampled data when we query the API using the spreadsheet add on. This is undesirable for how we are using the data.
The total results that we were getting before being returned was about 1100 results. We have altered the date range of the query to be only 3 days whereas before we were querying since the start of the year.
Initially that worked and the results were no longer sampled. Then 24 hours later the data appears to be sampled again.
The documentation says the following regarding sampling for the free account:
Analytics Standard: 500k sessions at the view level for the date range you are using
We are not using our analytics that heavily so I cannot understand why we would have hit the 500K limit?
It is also not clear to me what a "View Level" is? Any help on this would be greatly appreciated.