Unable to load certain URL in IFrame.. eg: google.com - iframe

I cannot load some sites in iframe .. is there any other things which i have missed . can you explain me, how can i load those url's in iframe?
<iframe src="http://www.google.co.in">
</iframe>

The site might be sending an X-Frame-Options: header to prevent embedding in a frame (a common way to prevent "clickjacking").
I know the Google login pages do that. The developer console in Chrome should tell you if this (or anything else) is the case.

Related

Multiple links to same page google analytics

So in google analytics I am seeing two links for every post I have on my wordpress site. See: https://www.screencast.com/t/JVoudis1eK8P
All the duplicate links are the same except for /404/ at the beginning. I checked my database and don't see these links. They do link to the correct page and don't display the /404/ in the url toolbar when clicked.
I inherited this site so I didn't set it up. Has anyone ever seen this? Is a plugin doing this? (We use yoast).
Thanks,
Mike
My suspicion isn't Google Analytics, but something on your site isn't right. Load up an Inspect session in your browser such as Google Chrome, and then go to the Network tab, and watch each request. You'll look for the "status" column for anything that's not a 200. I'd bet something like a JS, CSS, or image isn't loading, and your site is redirecting each 404 request to a page and passing the referring URL to it.

Making your site shareable on LinkedIn

I'm having a few issues with making our site shareable on linked in and I'm at a loss. The og: meta tags all look fine, the facebook scraper picks it up fine, but the linkedIn scraper does not... and the img etc are not on a protected folder or anything like that.
When inspecting the developer tools the get request to the url-preview?url= link shows that the img etc.. aren't there.
The image is less than 1mb, all og: meta tags are obeyed. The only think that may not be 100% is the image ratio is not 1/4 or 4/1 (it's 2/1)... But that is only a recommendation and not a hard and fast rule.
Does LinkedIn provide something similar to FB (https://developers.facebook.com/tools/debug/) where you can test the scraper and re-run it? Or is there another way to debug this? Any help appreciated.
https://www.hipla.co.uk (is the page i'm trying to share).
cheers
It transpires linked in doesn't offer a similar facility to FB or twitter to test the OG meta tags and re-scrape the page. They cache a page for 7 days and then re-scrape again. However, you can refresh the linkedIn crawler cache simply by appending GET params to the URL, i.e. https://www.hipla.co.uk?123.
I eventually figured out what our issue was. We were using a wildcard cert (for multi domain, so we could have a single ssl cert for multiple subdomains) which meant we had to set the server name in the apache default-ssl.conf file, but we had a typo in it for the www instance ... which meant it gave an SSL error (for the linkedIn crawler) which isn't debuggable (if that's a word) using linkedIn but was spotted as we got an SSL error when testing the twitter metadata tags using the twitter card validator. Hope this helps anyone else who has a typo in their ssl settings. Note that the ssl error was not visible using a browser(s) as all looked fine.

Site Hijacking RSS Feed and Entire Site in iFrame

The following site appears to be hijacking a client's content.
http://mothernova2.rssing.com/chan-24556607/latest.php
This is my client's site.
http://www.mothernova.com/
How would I go about blocking that domain from accessing the site? It also appears they are pulling the site into an iframe allowing full browsing.
FYI, the site is using WordPress, WordFence and iThemes Security (if there are any settings I should add for blocking).
You need to use a framekilling script, which uses javascript to check if your script is the top one. Here's a simple version:
<script type="text/javascript">
if(top != self) top.location.replace(location);
</script>
One drawback to this approach: if there is a legitimate site iframing your code, you need to check the referrer and start adding exceptions.
And a question to answer before you do it: you're getting a pageview and ad impression from the annoying framing site; is there any reason why you need to go to the bother, when they're sending a few viewers to your client's content?
The site owners of rssing.com are well known scrapers. And they are grabbing your content by RSS, hence the name rssing.com.
You can use the contact form to ask that they take your content down. Tell them they are clearly violating your TOS and copyright for your content.
(I had to do this in the past for my own content scraped from my site; they did remove my site at my request.)
Maybe I wasn't implementing the above suggestions correctly (I was adding them at the page level), but they weren't working for me. I did find this post and it seems to work as outlined.
http://forum.ait-pro.com/forums/topic/rssing-com-good-or-bad/
I updated my .htaccess file with the suggested code.
Brett

Hosting certain pages on a different domain in an iframe produces weird results

Folks,
I am working on an app that hosts certain pages from a different site in an iframe. Because the site is on a different network, I get prompted to log into their network in order to show the pages. All that is fine (the users of this site will be on the same network, so I'm not worrying about their logging in.)
However, on a different page, same app, I'm trying to do the same thing with a different set of pages from that same site. This time, I don't get prompted to log in, but do get the DNS Error page in the iFrame (I'm doing this in IE9, if it matters.) If I click "return to previous page" on the DNS Error page or right-click and click Back, the home page of the app (on the different network) appears in the iframe. After that, trying to load the desired page in the iframe works. But I'm baffled as to why and would prefer it load the first time I try it!
I'm guessing that the conjunction of the hosted pages being secure (HTTPS), their being on a different network, the iframe, and possibly also IE9 are somehow causing this. Anybody have any ideas?
Thanks,
Ann L.
If you go directly to the url that the iframe is pointing at do you get that same error or does it work? And make sure you go to the page the iframe is looking at and not what you think it is looking at. It may be that at some point in your process the page is generating the wrong url for the iframe.
The other thing is that it may be worth using fiddler to check your traffic to see if these other pages are trying to do some kind of redirect on you - they may be checking referrer and not liking connections from outside their network or similar.

capture details from external web page

I'm wondering if it's possible to capture details from the web page that a user previously visited, if my page was not linked from it?
What I am trying to achieve is to allow users to my site to find a page they like while browsing the web, and then navigate to a page on my site via a bookmark, which will add the URL (and possibly some other details like the page title) to a form which they can then submit to my site to add the page to a list of favourites there.
I am not really sure where to start looking for this. I wondered if I could use http referrer, but think this may only work if there is a link to my page?
Alternatively, I am open to other suggestions as to how I could capture this data - a Firefox plugin? A page which users browse other sites in an iframe, with a skinny frame on top?
Thanks in advance for your suggestions.
Features like this are typically not allowed by browsers for security and privacy reasons. The IFrame would work, but this is a common hacking technique so it may be likely to break or be flagged in the future.
The firefox addon is the best solution, but requires users to install it manually.
Also, a bookmarklet could be used. While they are actively on the target page, the bookmarklet could send you the URL.
This example bookmarklet would create a tinyURL for the destination page. You could add it to your database or whatnot.
javascript:void(window.open('http://tinyurl.com/create.php?url='+document.location.href));
If some other site links to yours and the user clicked on that link which took them to your site you can access the "referrer" from the http headers. How you get a hold of the HTTP headers is language / framework specific. In .NET you would use the Request.UrlReferrer; other frameworks would probably handle it differently.
EDIT: After reading your question again, my guess would be what you're looking for is some sort of browser plugin. If I understand correctly, you want to give your clients the ability to bookmark a site, while they are on that site, which would somehow notify your site about the page they're viewing. The cleanest way to achieve this would be a browser plugin. You can also do FRAME tricks, like the Digg bar.

Resources