I'm having a few issues with making our site shareable on linked in and I'm at a loss. The og: meta tags all look fine, the facebook scraper picks it up fine, but the linkedIn scraper does not... and the img etc are not on a protected folder or anything like that.
When inspecting the developer tools the get request to the url-preview?url= link shows that the img etc.. aren't there.
The image is less than 1mb, all og: meta tags are obeyed. The only think that may not be 100% is the image ratio is not 1/4 or 4/1 (it's 2/1)... But that is only a recommendation and not a hard and fast rule.
Does LinkedIn provide something similar to FB (https://developers.facebook.com/tools/debug/) where you can test the scraper and re-run it? Or is there another way to debug this? Any help appreciated.
https://www.hipla.co.uk (is the page i'm trying to share).
cheers
It transpires linked in doesn't offer a similar facility to FB or twitter to test the OG meta tags and re-scrape the page. They cache a page for 7 days and then re-scrape again. However, you can refresh the linkedIn crawler cache simply by appending GET params to the URL, i.e. https://www.hipla.co.uk?123.
I eventually figured out what our issue was. We were using a wildcard cert (for multi domain, so we could have a single ssl cert for multiple subdomains) which meant we had to set the server name in the apache default-ssl.conf file, but we had a typo in it for the www instance ... which meant it gave an SSL error (for the linkedIn crawler) which isn't debuggable (if that's a word) using linkedIn but was spotted as we got an SSL error when testing the twitter metadata tags using the twitter card validator. Hope this helps anyone else who has a typo in their ssl settings. Note that the ssl error was not visible using a browser(s) as all looked fine.
Related
I have created a website using WordPress. After some days, Chrome is showing Deceptive site ahead warning. But Firefox, Opera, Edge don't show any warning.
I have added my site to Google Search Console. In Security Issues tab, it says No issues detected. But google chrome is still showing the warning.
How can I fix this issue? Is there a way to submit my site for a review?
UPDATE:
I followed this link provided by #scytale and followed the following steps. But it is still the warning.
Scan Website for Malware or Threat: I have scanned my website using Sucuri, ReScan, VirusTotal, and Quttera. None of the scans detected any malware in the website.
Install SSL Certificate Properly: I am using AutoSSL from cPanel Inc. If I click on the alert icon at the left of warning, it shows the certificate is valid. So I guess my SSL is okay.
Redirect the Website From HTTP to HTTPS (301 Redirections): I have enabled in just like the example they provided. All http requests are being redirected to https.
Remove Mixed Content Warning: I have tested each individual page and haven't found any Mixed content.
The issue is, google search console is not showing any warning. So there is not option to request a review. Is there any other way I can ask google for a review? What else can I do?
According to Google the warning appears when "The site you try to visit might be a phishing site." However, if this were the case I would have thought it would be identified in Google Search Console.
dZone (a reputable security site) says the warning can appear for these reasons:
your website contains malware i.e. you've been hacked. However as
Google and other browser don't report issue this is unlikely.
issues with your site's SSL certificate (for HTTPS requests) maybe
installed incorrectly or not issued by an "approved" certificate
authority
Website traffic is not redirected to HTTPS (for fix see dZone article)
Mixed Content warning i.e. your https page "includes" content
obtained by HTTP (no S) URLs e.g. images, CSS, your or 3rd party
scripts. However, I'm pretty sure some of my pages use 3rd party
http affiliate scripts and I don't get this warning.
To fix check these issues and correct accordingly.
Its some time since I've used search console/webmaster tools but if Google has completed indexing of all your pages I would have thought it would also identify issues like mixed content. N.B. https://example.com and http://example.com are treated as "different sites" and you have to select appropriately in console. This Chrome extension claims to identify causes of mixed content for you (I've not used and have no idea as to its reliability).
Edit following request from Zakaria
I've no personal experience of this but I've done a bit of Googling:
Does the warning also explain why site is "dangerous" e.g. does it mention Phishing?
If you provide your domain name in the question then someone maybe able to identify why. I may not have time to check quickly, but others might be better qualified than me to check and comment.
Do you trust the theme and plugins you are using on your site? Google them for issues.
Chrome safe browsing, as well as using downloaded lists of dodgy sites; also locally analyses the properties of each page you visit and if it thinks suspicious forwards to Google for further checking. In which case it may be possible to find out which part of the page or script Chrome finds suspicious. See this Chrome Help Page.
Enter your domain here to see if you are on a "Google Blacklist"
Is your domain name new to you? Perhaps it was used for Malwre/Phishing in the past?
I'm looking for an easy way to share through LinkedIn without all that hassle with OAuth 2.0 which I don't see required when I see other pages that use this kind of sharing (and they didn't required anything from - I can straight out share).
Straight to the issue:
this one works: https://www.linkedin.com/sharing/share-offsite/?url=https%3A%2F%2Frefair.me
this one doesn't: https://www.linkedin.com/sharing/share-offsite/?url=https%3A%2F%2Frefair.me%2Fjob%2F494
Seems like beyond main domain I can't get sharing working. For instance from other site a link that goes deeper and is still shareable: https://www.linkedin.com/shareArticle?mini=true&url=https://bulldogjob.pl/companies/jobs/2043-programista-java-warszawa-bms-sp-z-o-o&title=Programista+Java&summary=&source=https://bulldogjob.pl
I also tested with and without source and summary query params. Anyone had that issue?
LinkedIn uses the Open Graph protocol (http://ogp.me/) to determine how pages are shared in LinkedIn.
You may also use the LinkedIn Post Inspector (https://www.linkedin.com/post-inspector/) tool to debug how various pages would be shared in LinkedIn.
I decoded your URL so I could get a cleaner look...
https://www.linkedin.com/sharing/share-offsite/?url=https://refair.me/job/494
So, let's try to visit your URL: https://refair.me/job/494 . The webpage you are sharing DOES NOT LOAD.
Is your site down for everyone? Yes, your site is down for everyone.
In order to share a URL on LinkedIn, you must fulfill the following minimum requirements:
The URL must load.
If you just want to test out the API, try using wikipedia.org or google.com as test pages.
Surprisingly, the old refair.me URL by itself works fine in LinkedIn, but that could be from some internal cache, from way back in the day when the page once did work. It certainly does not do so anymore.
For some reason sharing links on LinkedIn from my client's site does not work.
I've checked the specifications on https://www.linkedin.com/help/linkedin/answer/46687/making-your-website-shareable-on-linkedin?lang=en and it looks like all the og meta tags re correct. Sharing on other social media works without problems. It's only on LinkedIn where the Open Graph data is not picked up.
Here is a sample URL which does not work on LinkedIn:
https://nomadandvillager.com/bestemmingen/kenia/mbara/vrouwenbesnijdenis/
Am I missing something?
Would you be able to share a bit more on how you're getting the tags to render on Facebook and the like?
I encountered the same problem (My site runs on Angular), but my approach was to redirect the LinkedInBot via .htaccess.
Eventually, I gave up, and wrote a custom share button with LinkedIn's rest API
With it, you will be able to specify the details of the share, and avoid the messy workarounds for LinkedIn.
Direct Sharing on LinkedIn
Sharing via REST API
Original link: http://blog.crazy.technology/post/Clash-of-Clans---How-to-use-the-Web-API-570dd2b2
Two days ago we posted a new blog on a site with the aim of being picked up for the search term "live comedy in chippenham". It’s been indexed by Google and we’re now 2nd in the results for the search query. The bad news is that for some reason the post has been indexed as a https URL so all browsers give a warning when the link is clicked.
Firefox gives this error:
The owner of www.neeld.co.uk has configured their website improperly. To protect your information from being stolen, Firefox has not connected to this website.
The host has confirmed that it's not a server config error and we have other posts and pages on the site that are being indexed correctly. We're using WordPress and the Yoast plugin. I can't see anywhere in Webmaster Tools that could be causing the problem.
Can anyone offer any advice please? If you search Google for "live comedy in chippenham" you'll see the issue (it's the link https://www.neeld.co.uk/live-comedy-in-chippenham/)?
It's a really strange one but something I've experienced before.
It has mostly likely been caused by an external link to the page using https protocol which Google has followed before indexing the page. Google are very keen to index https pages at the moment so we might start seeing this kind of issue more often.
There's not a lot you can do other than wait for Google to realise their mistake and list the correct URL in the SERPS. You can help speed this along with a canonical link (which I can see is there), XML sitemap (which you've got) and a server level redirect of https to http.
Do not try to remove the page in Webmaster Tools as this won't have the desired effect and will stop Google reindexing the page properly.
Hope this helps.
First time posting so please bear with me.
I'm the unofficial web guy at the company I work for and I helped create our basic static HTML site.
Any work that I do to the site offline and then FTP shows up instantly on my machine. I rarely, if ever, need to clear the cache for changes to show up. However, within the company I work for, nearly half of the users never see the updates. Some do, some don't.
On the machines that don't I've cleared the cache in browser and through the internet control panel settings. Nothing. Still shows the stale content. The only thing that works - and I've seen this both in Chrome and IE is that when I add www in front of the URL is then shows the refreshed site. No big deal, right? Well for users who type in mysite.com without the in front will not see the updates. People who have favorited it like that, will not see the updates.
Now, on to what I've tried to fix it. After much research many people have steered me away from meta tag refresh so I haven't tried that, however, with the help of the IT guy we have, from what we can tell, set the HTTP header of the site to always refresh. This did not do anything for us.
I've tried changing image names in the HTML page when updating a photo and that didn't work either.
I haven't been able to find a .htaccess file so can I create one? If we (IT guy and I) changed the HTTP Header setting to always refresh but there is not .htaccess file will there be no change?
Any help or suggestions would be greatly appreciated.
I have searched on here for the answer and the two most suggested changes are HTTP Header and Meta refresh. HTTP header didn't help and it seems the Meta tag route is bad form.
This is a DNS issue. You need to ask the provider of your web services to add an A or a CNAME record for the domain's root.
If you don't understand the above, just call the provider of your web presence (the company that hosts your web server) and tell them you want yourdomain.com and www.yourdomain.com to go to the same place.