Access to sitemap? - asp.net

i created a site map with the name "Web.sitemap" in the root folder, and i need to feed this to google keywords. Any idea how i can access this file? I tried (domain)/Web.sitemap , but it doesn't load.
What is the proper way to access this file?
Thanks

Web.sitemap is typically used by the Sitemap control in ASP.NET to render menus and what not. It is not exposed publically, and in fact the default IIS configuration will block it from being loaded through the browser.
You may be thinking of a sitemap.xml file, which is an XML description of every page on your site used by search engines and crawlers. More information on this can be obtained from http://www.sitemaps.org/protocol.php

Not sure what you mean by feeding it to "Google keywords"? But if you want to submit a sitemap to Google Webmaster Tools (and search engines in general), it is an XML sitemap following the XML sitemaps protocol you want (as Mike wrote)

Related

How to track a PDF view (not click) on my website using Google Tag Manager

How can I track that someone visited the following URL of my website http://www.website.com/mypdf.pdf.
I tried using a Page View trigger on a Page View tag. I'm completely new at Google Analytics so not sure how to proceed. Most people are going to be going to that pdf directly via URL, as there is no link to it on my website, but I really want to be able to track how many people view it.
Thanks in advance!
You cannot track PDF views with the help of GTM. GTM for web is a javascript injector, and one cannot inject Javascript into a PDF document from the browser.
One way to circumvent this is to have a gateway page, i.e. have the click go to a HTML page that counts the view before redirecting to the document in question (naturally you could use GTM in that page). Since people go directly to the PDF URL this would require a bit of scripting - you would have to redirect all PDF links to your gateway page via a server directive, count the view and then have the page load the respective document.
Another even more roundabout way would be to parse your server log files and send PDF requests to GA via the measurement protocol (actually many servers allow to have log writes redirected to another script, so you could do this in realtime). I would not really recommend that approach - it's technologically interesting, but probably more effort than it is worth.
The short version is, if you are not comfortable fiddling a little with your server setup you will probably not be able to track pdf views. GTM does not work on PDF files.
Facing same issue…
My solution was to use url shortener (like bitly.com) which includes opening statistics.
Not the perfect solution but it works for direct pdf access from external source (outside your site).

How can I Checks for all “Not Secure” Pages?

I would like to retrieve the url from all the unsecured pages on my site.
Are there tools for this?
If I've understood well the question, you need to check, by crowling your entire site, for pages alerting the user about "security" issues.
In substance you got to find if any of your site's pages contains an "HTTP:" (not HTTPS) url inside the html code, while browsing yout site on https://yoursite.com.
If this is the case, i will use wget (to locally download the "rendered" pages of the site locally) and grep (to find any http:// link inside the code) for this task, under linux.
If the site is not generated by a CMS, and you can have access to the sources, I will parse the source files to search for any "HTTP:" inside the code.
BTW, this method does not help you in checking for linked objects (images, scripts...) using unsecure SHA-1 certificates or expired ones even under https://...
Could you kindly further qualify your specific needs?
Best regards.

Kentico Google Analytics page view

I'm looking at the GA traffic and I'm seeing page views for pages like this: /cms/getdoc/2d22c1db-ae83...angobjectlifetime=request
Is this page used when a user is viewing a document (PDF, Word, etc.)?
Not necessary. It could be any page within the content tree (including files). The cms prefix means it requires authentication (it's usually within the administration interface) and the getdoc handler means the url is permanent one (uses the GUID that follows), so you always get this page/file no matter where in the content tree it is (after you move it for example)
Google Analytics records anything hit to your website which is used to access either to access a legitimate page or used to access a resource on your website. You should use filters during report to filter out this data.

SiteMap for the search engines

I have to make an api call and get the response. This response contains more than 4000 urls.
I have to list all these urls in the sitemap for the search engines to crawl easily. I have to write a handler for doing this task. Can someone suggest me an example for doing this.
I'll assume you are talking about a sitemap in XML format, but you didn't specify what the source is besides that you are to do an API call. However, the 3rd or so result from a Google search on "asp.net google sitemap" should give you a perfect starting point:
http://www.mikesdotnetting.com/Article/94/Create-a-Google-Site-Map-with-ASP.NET
I would suggest creating an ASHX handler (File -> New -> Generic Handler in Visual Studio) instead of a page like they do in the example.
Upload the handler to the website and add the sitemap to e.g. Google by using their Webmaster Tools.
A quick search on Google resulted in this link:
XML sitemap with ASP.NET
Which should get you most of the way with the handler and composing the XML.

How and where to add a robots.txt file to an ASP.net web application?

I am using ASP.net with C#.
To increase the searchability of my site in Google, I have searched & found out that I can do it by using my robots.txt, but I really don't have any idea how to create it and where can I place my tag like ASP.net, C# in my txt file.
Also, please let me know the necessary steps to include it in my application.
robots.txt is a text file in the root folder that sets certain rules for the search robots, mainly which folders to access and what not. You can read more about it here: http://www.robotstxt.org/robotstxt.html
The robots.txt file is placed at the root of your website and is used to control where search spiders are allowed to go, e.g., you may not want them in your /js folder. As usual, wikipedia has a great write up
I think you may find SiteMaps more useful though. This is an XML file which you produce representing the content of your site. You then push this to the main search engines. Although started by Google all the main search engines have now agreed to follow a standard schema.
Increasing your Google score, and SEO in general, isn't something I've know much about. It sounds like a black art to me :) Check out the IIS SEO Toolkit though, it may offer some pointers.
Most search engines will index your site unless a robots.txt tells it not to. In other words, robots.txt is generally used to exclude robots from your site.

Resources