Find #pages in website - google-analytics

I'm collecting data on complexities of several domains- represented by total pages, visited and unvisited.
I was initially finding what I wanted from Google Analytics by drilling down to Behavior -> Site Content -> Landing Pages but wasn't sure if that was returning unvisited sites. Then I tried All Pages per domain, but that returned like 1,800 results for "pages", with params in some cases /Pages/Results.aspx?k=update.
That being said, I don't think I can rely on GA for total pages per site.
Then I thought about using a web scraper, namely web2disk or httrack.com, to scrape for the #pages per domain. Is that a good path to take? Is it necessary to get this information?
Thanks

If you want to know how many pages there are on your site you need to crawl your site to find all the pages. Because of the way it works Google analytics will 100% only show you data on pages which have been loaded (which fires the analytics code) in a browser.
http://www.screamingfrog.co.uk/seo-spider/ is a paid for crawler you can use to find all the pages (£99), or you could potentially try to hack something together using a free crawler like http://import.io (disclaimer: I work at import.io) to get all the URls.
Find all visited pages via GA:
Behaviour -> Site Content -> Landing Pages does not give you any pages which were not 'Landed upon'.
Then I tried All Pages per domain, but that returned like 1,800 results for "pages", with params in some cases /Pages/Results.aspx?k=update.
To remove the params from the page URls you can use a report filter at the top right of the table. Click 'advanced', and use the tools there to exclude params from URls.
Alternatively you can switch your primary dimension to 'Page title' if you have Unique page titles for each page (and identical ones for pages with params).

Related

Google Analytics - Some weird URLs are being registered in behavior reports

We made a google analytics account for one of our clients as part of the requirements.
Under Reporting tab, we have sections like Behavior -> Site Content -> All pages. In All Pages section we are able to see a table which contains all the urls which are viewed with respective pageviews , unique pageviews and other dimensions.
A normal page view looks like:"/pwsportal/faces/homePageNav/mktplan_adf.Ctrl_9_afrLoop_1234423".
Some how there are some weird page urls like :
I tried using Exclude Filters and couldn't eliminate these kind of urls.
From one on the blog i got to know that if any url contains any script tags it is a part of hacking technique called cross site scripting.
Finally i am here to find a solution to eliminate the these kind of existing urls and to prevent them from getting registered in future into google analytics.

Website doesn't appear In first 20 list of links in a Google Search

I have recently completed one website and am now working on SEO for same, in order to get listed in first 20 links of Google search.
Basically our website is on car rent services in Goa, India.
I completed with following things in SEO
Verified my website in Search console
Uploaded the .html file on server
Also uploaded robots.txt.
Generated site map and uploaded the same on server
Pages submitted for indexing ....in fetch as Google some pages show status as complete and partial
Key words added to pages using plugin.
All the above has been done but still our website is not getting listed in first 20 of Google search.
Pages submitted for indexing but still count shows 0.
Please help
Thank you
SEO is a slow and steady process; naturally it will take time to achieve the top position in search result.
Things you should do;
1) Add meta title and description to each page based on the keywords
2) Include keywords in URL of the page
3) Create a webmaster tool and use fetch page option to get listed in search engines (Example: Google webmaster is must needed)
4) Create brand pages in social media sites and interlink with website
5) submit your website to dmoz.org
https://moz.com/beginners-guide-to-seo --> Useful guide for you
Check out Google's Webmaster Guidelines: https://support.google.com/webmasters/answer/35769
Following these guidelines will help Google find, index, and rank your site.

How to make search engines crawl tens of thousands of pages in my website

I have a large amount of items that each item has a page like this site/item_show.aspx?id=The_ID_here there are tens of thousands of items and each day nearly two thousands are added. Furthermore each item has a description in its page so for each item, its page should be crawled by search engines.
My question is with this amount of data described: How can I generate sitemaps or anything like that to make all items visible by google and other search engines?
It is clear that I cannot show all items in the first pages but I can make pages that simply just contains the link of items and provide tens of them each page for just search engine. Would it work or is it anything better to do for making items be indexed by google?
essentially there are 3 methods which will help you with the Mass-Indexing :
1. Create an XML SiteMap for all of your pages and link to it from you HomePage.
2. you should have Google Webmaster Tools Installed and you can load that same XML file into it.
3. Have an Organized Categories structure - depending on the type of your site think about a logical Categories Structure for example in eCommerce stores all the products are categorized by product Main-Category then a product Sub-category and sometimes by Brands etc... - of course you should do this via your Shopping-cart platform - Just remember that if you begin chaging the URLs-structure you'll need to take care of all the Redirects from the old URLS to the new ones.
First, use XML sitemaps and submit those to Google (note that i said sitemapS - more than one).
Next, ensure that your on-site content is nicely organised into categories and sub-categories - ideally you'd want all elements to be reachable in as few clicks as possible without users (or Googlebot) having to resort to the search function.
Finally, ensure that your more popular / important items are featured in the homepage or 1-2 clicks deep, and get links and social shares to those specific product pages.
Be popular and get links to your site. Have a good server which can handle the crawl.
There is also not a hard limit on our crawl. The best way to think
about it is that the number of pages that we crawl is roughly
proportional to your PageRank. So if you have a lot of incoming links
on your root page, we’ll definitely crawl that. Then your root page
may link to other pages, and those will get PageRank and we’ll crawl
those as well. As you get deeper and deeper in your site, however,
PageRank tends to decline...
https://www.stonetemple.com/matt-cutts-interviewed-by-eric-enge-2/

How to track conversion funnel in Google Analytics where a banner ad is shown on many pages.

I have a website that features a call to action/promotion button on nearly all pages of the site.
I have currently configured a conversion funnel that shows me how many people arrive on the call to action page, and then how many people make it to successfully complete the action page.
What I want to see though is how many unique visitors over the reporting see the banner at the top of the funnel.
eg. Something like this:
Visitors accessing website: 1000
Visitors clicked on call to action page: 100
Visitors successfully submitted call to action form: 45
My initial thoughts was to do this using the frontpage only, but I forgot that this banner/call to action ad is featured on many pages around the website. Many people find the site through SEO and never even pass through the frontpage.
Is it possible to use a wildcard for a domain or something similar in Google Analytics? Or maybe I am approaching this the wrong way.
Last of all - I know I can accomplish this by pulling up 2 reports: site wide unique visitors and comparing that to how many people hit the first stage of the existing conversion funnel. But it's a hassle to have to do this regularly manually.
While using funnel analysis, it is normal to have funnel steps that represent more than 1 urls. Take the basic case of ecommerce sites, where the final goal maybe the same transaction completion page, but the funnel step corresponding to product page can be triggered by many different product pages and not just one.
Based on the page url structure of your website, you can choose any of the below 2 match types to add multiple urls to a single step:
1, Begins with : If all the different pages displaying the ad have a set of common characters in the beginning, then use this.
2, Regular Expression Match : If the different pages that contain your banner ad how totally unrelated url, then find a suitable regex that can capture all those urls

Google Analytics referral triggered by a bookmarklet

I have a question regarding Google Analytics and unwanted referral stats generated by a bookmarklet.
I have a web service with GA installed. My users are using a bookmarklet to accomplish a certain task while visiting some other web page. Bookmarklet creates an iframe and opens up a page which is also on my domain and that page contains the same GA code.
For some reason GA sees those web sites (pages that bookmarklet was used on) as referral pages. That creates a problem for me since those pages are not real referrals (no actual links to my site). I have no desire to track pages my users marked with the bookmarklet.
It’s important to mention that bookmarklet page must be a part of the same domain as my main page. I can not move it on other domain or subdomain.
This is what I tried so far:
I’ve created a new GA account (subdomain.mydomain.com) and used it only on my bookmarklet page hoping that all stats related with the bookmarklet will appear on that account. This worked only partially. Stats for the bookmarklet started to appear on the new account but my original GA account continued to track referral pages.
We tried to use a pop up window to load a web page instead of the iframe. No difference.
Any help on how to get rid of unwanted referral sites would be appreciated.
See _setReferrerOveride:
_setReferrerOverride()
_setReferrerOverride(newReferrerUrl)
Sets the referrer URL used to determine campaign tracking values. Use this method to allow gadgets within an iFrame to track referrals correctly. By default, campaign tracking uses the document.referrer property to determine the referrer URL, which is passed in the utmr parameter of the GIF request. However, you can over-ride this parameter with your own value. For example, if you set the new referrer to http://www.google.com/search?hl=en&q=hats, the campaign cookie stores a new campaign with source=google, medium=organic, and keyword=hats.
_gaq.push(['_setReferrerOverride', 'URL-YOU-WANT-AS-REFERRER']);
Or, you could try
_addIgnoredRef():
_addIgnoredRef()
_addIgnoredRef(newIgnoredReferrer)
Excludes a source as a referring site. Use this option when you want to set certain referring links as direct traffic, rather than as referring sites. For example, your company might own another domain that you want to track as direct traffic so that it does not show up on the "Referring Sites" reports. Requests from excluded referrals are still counted in your overall page view count.
Async Snippet (recommended)
_gaq.push(['_addIgnoredRef', 'www.sister-site.com']);
You would have to grab the referrer and populate it dynamically. Probably with parent.document.referrer Of course this might make any referrals (non-bookmarklet) from these sites not record in the future. And, at some point you would need to clear them.
The most simple solution, if you don't need to track the hits from the bookmarklet at all, is to simply not include the GA code in the web page when it is opened by the bookmarklet.
Your bookmarklet can open the page like http://yoursite.com/?mode=bookmarklet
And in your server side code you can use something like
if ( mode != "bookmarklet" ) {
outputGaCode()
}

Resources