Hi I tried searching for information regarding this but get bombarded with other search results, might not be putting in the right question not sure….
Anyways hopefully someone here can help, what I’m having an issue with is my site which has a mobile version that’s separate from the main site, the mobile version can be used by anyone and has a link from the main one but anyone with a mobile device gets redirected to the mobile one. The mobile site is blind to search engines using the nofollow tag on any links that lead to it and hasn’t been cached which is what I want because it would be considered duplicate content, anyone with mobile devices gets redirected to equivalent page on mobile site.
I previously forgot to remove the analytics line from one of these pages and noticed it got cached in the search results so I removed it, which is the issue, how do I know how many people were visiting here (mobile) now etc…. Is there a way to get analytics information without having it go to the search results for duplicate content?
This seems to be a problem that can be fixed with Google Webmaster Tools, or tagging the content with a rel="canonical", please take a look at:
https://support.google.com/webmasters/answer/66359?hl=en
Related
On my Wordpress site which is affiliating with Amazon associate, i have many native shopping ads for products, that normally displays the products pictures with their names and the prices.
Lately, i activated ublock origin on my browser, then i realized there is only a white space in place of the products, at first i thought it was only due to slow internet, then i said maybe it is the cache so i deleted the cache, but nothing changes .... it took me a while to realize that ublock is killing all my native shopping ads.
which means this was going on for a while and i never noticed. maybe that is why the low click through, even that i have a good number of visits!
without adblock, my site look like this:
enter image description here
after adblock, it looks like this:
enter image description here
how can i fix this? is there any workaround it?
I searched in google and i found about asking visitors to disable adblock. that won't do, because it makes them suspicious of the site and usually just leave it.
I know i can put a picture of the products then write information on it with affiliate text link, but that won't give me the same design and usability as amazon's native shopping ads.
So, is there any way that i can make my amazon native shopping ads visible even with adblocking activated? some WordPress plugin or something?
Your help is highly appreciated, I really need this.
Thank you.
No, there isn't an easy way to do this. Those are ads and users with adblockers want ads blocked.
You can try building your own system that delivers the ads and hope that adblockers do not adjust their filters, but that'll probably mean playing whack-a-mole: you change things, they adapt, you change them again, they adapt again.
I've got a site that has multiple share buttons on entries in a WordPress site.
We designed this so there are no individual entries to view, they're Podcasts and videos. The listing page has a minimum of 10 entries, each with share buttons.
Currently the share links and titles are working correctly. But the page is not recognizing the og:image, and instead is picking up the default logo for the site itself.
I read another post on Stack Overflow that said it might be an issue for LinkedIn if the image is utilizing SSL for the link. But I just find that hard to believe.
The other issue I'm struggling with, the docs say once an image is scraped it stays cached for approximately 7 days.
I had an issue with FaceBook and there's a debugger that allows you to rescrape the page which let's me verify my changes worked.
My two questions are, is there something other than the og:image i should be specifying? since I can't specify it per post, it's in the head of the page itself, i would think it would pick that up. No?
Second, is there a way a developer can re-check after the meta info has been changed to see if the changes worked, without having to wait the TTL on the cache?
try this:
url/link?blah=1
url/link?blah=2
url/link?blah=3
to get around the cache.
This should trick it into thinking its a new page each time.
Can i get a link to test?
Anthony Walz posted the correct answer. Through email he also helped another problem i had which corrected a new issue i didn't realize I had until i looked.
my LinkedIn shares were not picking up the show title, they were picking up the page description instead (i have several podcasts showing on one page, we don't use individual post pages, they all play from the listing.)
he pointed me to the developer docs on formatting sharing links
Which gives a real world example - here:
https://www.linkedin.com/shareArticle?mini=true&url=http://developer.linkedin.com&title=LinkedIn%20Developer%20Network
&summary=My%20favorite%20developer%20program&source=LinkedIn
Thanks a ton for assist Anthony!
Recently ran into some trouble with Google DFP that I'm hoping others have had.
We have a site that's served via SSL and it contains some Google DFP ad tags. Google DFP's debugging console shows no errors in the tags or our implementation of them. (i.e. the tags themselves are fine)
However, the ads are getting served via different methods. Some of the iframes get served as FriendlyFrames and some get served as SafeFrames. The SafeFrame ads appear correctly. The FriendlyFrame ads don't show up.
It appears that the FriendlyFrame ads are running afoul of some sort of browser security measure (likely because the pages are served via SSL).
I looked into this in the DFP docs but haven't found anything that explains how to solve the issue. There is a setForceSafeFrame method available that I've tried to use but it doesn't actually seem to do anything when I try to use it:
https://developers.google.com/doubleclick-gpt/reference#googletag.PassbackSlot_setForceSafeFrame
I've setup a test page demonstrating the issue here:
https://methnen.com/ad-test
There should be 5 separate ads on the page. If you get all of them refresh the page until you get at least one ad that doesn't show. The broken ads are being served as a FriendlyFrames.
Really hoping someone knows what the heck is going on.
FYI and possibly helpful for anyone else who might run into this at a later date:
Turns out the Ad Ops person hadn't set things up on their end to have enough inventory to fill all of the slots and there was nothing wrong with the tagging at all. The empty FriendlyFrames are apparently what DFP serves up when it decides it doesn't have anything to fill a given slot.
Try to force render all ads in SafeFrame
googletag.pubads().setForceSafeFrame(true);
More about it here https://developers.google.com/doubleclick-gpt/reference#googletag.PassbackSlot_setForceSafeFrame
Is it possible to track if someone links to data on my site? Specifically if my data is used in a site dynamically generated by a developer program? I would like to know if someone is blatantly passing off my site's data as their own. There are obviously ways around directly linking to content, such as content manipulation or even manual manipulation. But if someone where to link(or directly add word for word or manipulate) my content into their website, is there a way to track it?
Can I avoid someone being able to scrape my website at all, or is everything just up for grabs?
the best answer and the easy one is called GOOGLE - WEBMASTER TOOLS!
HERE
actually doing that is very hard and you would need to crawl the web to discover those links that address to your pages... dynamic content as well is linked so it would be find by google as well.
this tool will allow you to see outer links that address to your site.. and you can check them.
for extra - you can monitor requests and traffic to your site and find ip's that are using the same page over and over again. that can tell u that an outer page is dynamically loading content from your web page.
EDIT:
here is a good article in this subject: link - scroll down and you can see the use of google
webmaster tool with some other progrmas and method.
here is a good start guide to the google webmaster: link
ENJOY!
Basically a client has asked for his WordPress website to be turned into a mobile website as well. I have never attempted this and know nothing about SEO.
However the issue has arisen that this may cause duplicate content issues with Google, and therefore both sites may be dropped in the rankings.
I was looking at turning the website into a mobile site via one of the available WordPress mobile website plugins.
My question is if duplicate content will be an issue? Has anyone ever tried this?
After doing some reading I kind of think it may be possible to tell Google not to index the mobile website, although as I understand it It would be the same set of files. So I am unsure that if I tell it not to index one of them, that it will drop the other one as well.
Can anyone with WordPress and SEO knowledge clear this up for me?
In my opinion, if you have two indexed URLs with same text, there is duplicate content. And for Google, duplicate content is always an issue.
If you decide to no to index the mobile version of site, there will be not duplicate content because only one version of site will be indexed by Google.
Duplicate content is independent of your CMS (Wordpress), it's just a story about two indexed pages with same text.