I lost all the likes on my website on Wordpress then I bought the domain. It turns out that is the same site, but now no longer use the wordpress.com but .com (http://sobreasdeliciasdavida.com/).
Despite recent, my blog already had good statistics and the loss of more than 500 shares in Facebook brings my blog back to its beginning.
Can you offer the option of importing the likes to the new domain since the posts are the same?
Is there any way to do this?
Oftentimes when you move a well-established site, you'll want to set up a 301 redirect from the previous site. It's a permanent redirect that ensures that people following links to your previous site end up at your new one. I should point out, though, that your blog is far from taken back to its beginning. Remember, content is king, and you now have a site that's totally under your control and is already packed with great content, content that you know people respond to, like in social media venues, comment on, etc., etc. Don't worry about the 500 you might not get back because you certainly have thousands more on the way if you just keep doing what you're doing.
If you are directly using the facebook code in your website, then you can check this out. http://searchenginewatch.com/sew/how-to/2172926/maintain-social-shares-site-migration
Related
I'm working with WP on different subdomains, and for some reason a link on the landing page isn't going to the subdomain, but rather it's trying to go to a page that doesn't exist on the main domain level. I'm guessing this is some kind of auto direct issue, but the link is correct and I'm not sure what is causing this.
The main domain is in staging status: staging2.definingstudios.com. There are three links there, Lifestyle, Schools, and Commercial. The linnk to schools had been doing this too, but then it stopped and is working properly, but the Lifestyle one is trying to go elsewhere.
Thoughts?
Thanks,
Christine
This isn't a WP issue but generally an issue with the theme. They tend to save this data in the DB. A lot of the time its serialized and doesn't get updated.
You can do a few things. Search the DB and see if you can find it. Use a plugin to do it or 95% of the time just go into that page and update/save and it generally fixes it.
I'm not a really advanced Analytics user, so I've been trying to Google this, but haven't come up with a great answer. My analytics says 95% of my site visits to my blog today have come from site38.social-buttons.com and yesterday it was another subdomain of the same site. I visited social-buttons.com, but am unfamiliar with it, and have never deliberately put that code into my Wordpress site. I do have some plug-ins installed, which are "Subscribe / Connect / Follow Widget", which displays my social media links, and also "Really simple Facebook Twitter share buttons", which puts the like links on my posts.
My questions are, how are people finding my site through social-buttons.com? And are these quality hits?
Thanks, I appreciate any info!
This kind of visits are called Ghost Referrer Spam since they never reach your site. They use a GA weakness to make a fake visit and get a record in your data.
They do it to get traffic, people get curious to see who is visiting them and click on the link.
This specific Referrer Spam is nasty because it make multiple visits at the same time, is related to the number of the subdomain so if it says site38... it hits with 38 visits, I've also have many of these, here is a screenshot I took:
In my case is a different simple-share-buttons.com but is the same thing.
The easiest way to stop it is by making a filter for each spammer in your GA. Check this article to find more detailed information http://www.ohow.co/block-social-buttons-simple-share-buttons-referral/
As an alternative, you can make a more general filter to take care once and for all of all the Spammers by making a list of Valid Hostnames, this is more advanced and you have to be more careful. You can find more information about this solution here https://stackoverflow.com/a/28354319/3197362
It's actually referral spam. Take a look at this https://www.mooresoftwareservices.com/Web-Commerce/social-buttons-com-referrer-spam
So unfortunately they are not good quality hits.
Hypothetical Situation: I have a small obscure website called "miniatureBoltsInCarburetors.com" which provides content about the miniature bolts which hold a carburetor together as well as some general related automotive information. My site also has a single page which allows someone to find the missing bolt in their carburetor, and while no one will access this page directly from my website, one billion other popular automotive sites have embedded this single page in their website using an iframe, yet not included a link back to my site.
I recognize that this question is related to SEO which is considered off topic, however, all of the many SEO related forums discuss the marketing steps one could take, and not the programming steps or strategies, and hope others will allow this question to be answered here.
I wish my site "miniatureBoltsInCarburetors.com" to be ranked high for general automotive searches. What could I do to allow the 3rd party sites which include an iframe back to my site to improve my ranking? Could using JavaScript in the iframe to create a link on the parent page provide any value? What about when my server renders the page, use PHP to get the referring URL from $_SERVER, and include it in the content?
I am providing a solution here. Not sure if this is what you want though.
In your page which is used by other websites in iframe you can put below Javascript. This javascript checks if the webpage is opened inside an iframe or directly in browser.
So using this check when you see it is opened in an iframe. On click on something navigate to your website.
// This works in all browsers
function inIframe () {
try {
return window.self !== window.top;
} catch () {
return true;
}
}
Also for your reference you can check the below URL.
How to prevent my site page to be loaded via 3rd party site frame of iFrame
Hope it helps.
Iframes are seen seperate pages by Google. Your approach may end up being penalized due to being sourced from untrusted site. According to Google Webmaster Support
Frames can cause problems for search engines because they don't
correspond to the conceptual model of the web. Google tries to
associate framed content with the page containing the frames, but we
don't guarantee that we will.
One of the best approaches to rank higher for a specific keyword is, make multiple related sites. In your case a 3-4 paged site about carburetors, bolts, other things your primary site contain would do it. These mini sites will be more intense about the subject due to less page count. Of course they should contain unique articles on each page. Then link from mini websites to primary websites and you can see the dramatic change.
In fact, the thing you are trying to do was a tactic to rank competitors down worked occasionally a few years ago. Now, it is still a risk.
I see. You don't want to mess up the page for your own site, but you want to do something with all the uncredited embeddings.
The solution is fairly simple:
Create a copy of the page.
Switch your site to use the copy.
Amend the version that countless other sites are embedding, so that there is a small link back to you. Or, add an iframe blocker script that will load your site.
If the page is active (ie user interacts with it to find the missing bolt) you could include a sales message with the response encouraging the user to visit your site.
I think that your goal is getting your link onto these other sites long enough to get indexed by Google before it is noticed by the people doing the embedding, so it's a bit of a balancing act.
I see conflicting advice about how Google indexes iframes. You should use a PageRank checker to see if the existing iframe page url has PageRank, and compare it to the page that you embed it on.
I dont Think you need to worry ,.
Google bot does seem to crawl through Iframes ,but the Web-Page Containing that Iframe is not Credited for that Content .. In other Words,, Page-Ranking of that particular Web-Page do not Change due to Contents from Iframe .
is IFrame crawled by Google?
Do robots crawl iframes?
I have a website that I made very quickly a while ago, using a WordPress theme. I completely forgot about it for a few months and checked the traffic for the first time today, and surprisingly it has been getting a lot of visitors and generated some income.
Currently the design is pretty horrible and I am 100% positive that if I re-design the website myself, I can get so many more visitors and conversions.
So I'm thinking about getting rid of WordPress and publish a new website using Bootstrap, and keep the same content and URL that I had.
But I'm scared that that would mess up my SEO and lose my organic rankings. I am on the first page of my main keyword and I would hate to lose this spot.
When a site goes through a design reconstruction, are there any specific steps that I should take? Should I just keep the WordPress site to be safe? Or am I worrying about something that won't even happen? I would love to hear any tips or feedback about this.
I have had recent experience of just this problem.
The URLs from the wordpress site are a commodity which is invested in search engine servers. Your ranking (which has taken time to accumulate) is in part dependant on preserving the URLs of the pages of your site.
You will ideally need to place a redirect (.htaccess file if using apache) from the old URL's to the new ones.
Rushing into commissioning a new site without reseaching this will cause you huge SEO loss that takes six months or more to recover from.
See this for more information.
Please take your time on this. I have too many companies call me on this when they screw it up.
Build out the new site on a test site that is not indexed, make sure it works, make sure the URL's are the same, test it.
Make sure you have a perfect htaccess file and I mean 100% perfect. Flip the new site on with the new and updated .hatccess file and the make sure your sitemap and robots.txt file are steller. Submit all of it to Google Webmaster and to Bing Webmaster Tools. If you change your URL's you are going back to the stone age. If you keep the URL's the same you will not see an issue.
-Matt
I have a self-hosted wordpress blog, and as almost expected, I found there's another blog scraping my contents, posting a perfect copy of my own posts (texts, images not hotlinked but fetched and reupped to the clone's server, html layout within the posts) with a few hours of delay.
however I must confess I'm infuriated to see that when I search Google for keywords relevant to my posts, the scraping clone always comes first.
So, here I am, open for suggestions, would you know how to prevent my site from being successfully scraped ?
Technical precisions :
the clone blog appears to be self-hosted, and so am I, I'm on a debian+webmin+virtualmin dedi
my RSS feed is already cut with a "read more on" halfway. Hey, I just thought I should publish a post while assigning it a date like 2001-01-01, and see if it appears on the clone blog, that would allow to know if my RSS is still used as a signal for "hey, it's scraping time !"
my logs can't find the scraper among legit traffic, either it's non-identifiable or else it's lost among the flood of legit traffic
I already htaccess-banned and iptables-banned the .com domain of the clone, my contents are still cloned nonetheless
the clone website makes use of reverse proxies, so I can't trace where it is hosted and what actual IPs should be blocked (well, unless I iptables-ignore-ban half of Europe to ban the whole IP ranges of its data storage facility, but I'm slightly reluctant to that !)
I'm confident this isn't hand-made, the cloning has been running for two years now, every day without fail
only my new posts are cloned, not the rest of my website (not the sidebars, not the wordpress pages as opposed to wordpress posts, not the single pages), so setting up a jail.html to log who opens it page won't work, no honey-potting
when my posts contain internal links pointing to another page of my website, the posts on the clone won't be rewritten and will still point to my own website
I'd love help and suggestions with this issue. Not being cloned, but losing traffic to that bot while I'm the original publisher.
You can't really stop them in the end, but you might be able to find them and mess with them. Try hiding the request IP in an HTML comment, or white-on-white text, or just somewhere out of the way, then see what IPs show up on the copies. You can also try to obfuscate that text if you want by turning it into a hex string or something so it's less obvious to someone who doesn't know or make it look like an error code, just so they don't catch on to what you're doing.
In the end, though, I'm not sure how much it will buy you. If they're really inattentive, rather than shutting them down and calling attention to the fact that you're onto them, you can feed them gibberish or whatever whenever one of their IPs crops up. That might be fun and it's not too hard to make a gibberish generator by putting sample texts into a Markov chain.
EDIT: Oh, and if pages aren't rewritten too much, you might be able to add some inline JS to make them link to you, if they don't strip that. Say, a banner that only shows up if they're not at your site, giving the original link to your articles and suggesting that people read that.
Are you willing to shut down your RSS Feed? if so you could do something like
function fb_disable_feed() {
wp_die( __('No feed available,please visit our homepage!') );
}
add_action('do_feed', 'fb_disable_feed', 1);
add_action('do_feed_rdf', 'fb_disable_feed', 1);
add_action('do_feed_rss', 'fb_disable_feed', 1);
add_action('do_feed_rss2', 'fb_disable_feed', 1);
add_action('do_feed_atom', 'fb_disable_feed', 1);
it means if you go to a feed page, it just returns with the message in wp_die() on line two. We use it for 'free' versions of our WP Software with an if-statement so they can't hook into their RSS feeds to link to their main website, it's an upsell opportunity for us, it works well is my point, haha.
Even though this is a little old of a post, I thought it would still be helpful for me to weigh in in case other people see the post and have the same question. Since you've eliminated the RSS feed from the mix and youre pretty confident it isnt a manual effort, then what you need to is better stop the bots they are using.
First, I would recommend banning proxy servers in your IPTables. You can get a list of known proxy server addresses from Maxmind. This should limit their ability to anonymize themselves.
Second, it would be great to make it harder for them to scrape. You could accomplish this in one of a couple of ways. You could render part, or all of your site in javascript. If nothing else, you could at least just render the links in javascript. This will make it significantly harder for them to scrape you. Alternatively, you can put your content within an iframe inside the pages. This will also make it somewhat harder to crawl and scrape.
All this said, if they really want your content they will pretty easily get by these traps. Honestly, fighting off webscrapers is an arms race. You cannot put any static trap in place to stop them, instead you have to continuously evolve your tactics.
For full disclosure, I am a co-founder of Distil Networks, and we offer an anti-scraping solution as a service.