I have an odd situation with a DEV site at work. Last week, our sys admin was experimenting with moving our DEV servers but didn't get chance to finish the job so rolled back his changes.
When I type in this URL http://mylocalsharepointsite/anypage.aspx then I get the page easily. However, when I change the protocol to https I first get a message warning me about the certificate, then when I click ignore warning I get a 404 error.
I know the site and pages are there because I go to IIS and can see the site, the directory it is mapped to and the SSL certificate, which since it's DEV we generated ourselves.
I suppose my question should be this, what are the things I should look at first, is there anything obvious I've overlooked.
The problem here was with IIS and the certificates. I re-generated the certificates, even though there didn't seem to be anything wrong with them.
This page helped me a lot, especially the two tools: SelfSSL and SSLDiag.
http://www.iis-aid.com/articles/how_to_guides/creating_self_signed_certificates_iis
Related
I shared a link with someone to a firebase site that I was hosting, and it worked for some time, but then all of a sudden they said they were getting the message:
We're sorry... ... but your computer or network may be sending
automated queries. To protect our users, we can't process your request
right now. See Google Help for more information.
I was also getting it, and started checking my other firebase hosted sites and started getting the message on all of them. I didn't understand. I couldn't find a common link to understand why it was happening. So many sites linked it to a reCAPTCHA problem, but my sites don't use reCAPTCHA...
I found this link:
GitHub forum Link
The user recommended making sure the url started with "https". As soon as I typed my url with "https://" at the start, everything came up. At that point, I tried all the other URLs, and they worked too. This may be a rule for all Google-related sites.
I'm not sure if it's relevant that Chrome often trims the url in the "omnibox" or address bar, hiding the protocol, making it easy to miss when copying/pasting? E.g. :
Note, I tried accessing these pages without https (by typing "http://") but my browser now seemed to correct it and force it to be "https://", so I couldn't replicate the problem again.
I don't know exactly why it started, but I know that I wish I found this information sooner, because it was very frustrating, and the info out there wasn't helpful, except for the link I posted above. So hopefully, when someone like me searches for "firebase" and the error text "your computer or network may be sending automated queries", they might see this and possibly be saved a headache.
The title pretty much says it all. When I'm running HTTP(S) Test Script Recorder, one particular page becomes unresponsive - when I click on something it just reloads. The recorder itself is working fine, it is recording every step. And the problem is not proxy related. I've successfully recorded other pages of the same website. When I'm not recording, there are no problems.
It's a .NET 3.5 project.
The page itself has a lot of forms, file uploads, etc, but as far as I know, it should not interfere with recording or even more - with browsing the page. When debugging the project, no breakpoints are hit, so I assume that something gets lost before reaching the server.
Browsers that I've tried: tried FF, Chrome, Edge, IE.
Tried recording the web locally and online. Same thing.
Played around (reinstalled and whatnot) with certificates, didn't help.
Has anyone encountered such a problem? What could be done to fix this? I'm more interested in finding the solution, than a way around (blazemeter, badboy). Any help would be very appreciated.
EDIT: I tried recording with blazemeter and it worked. But when looking at view results tree I noticed that the request path and parameters don't change, even when in the HTTP Request Sampler they are different. So there's no solution yet
This often happens to me and what I've found is that JMeter changes the root certificate in the bin folder every week. So usually the HTTP traffic is fine but certain HTTPS traffic won't work. So make sure that your browsers are seeing the latest JMeter certificate and not using an old one that doesn't exist anymore. On Windows, Chrome and IE use the certificates in Internet Options, while Firefox needs to have it added to it manually.
It turns out that in Test Script Recorder HTTP Sampler Settings choosing Type: Java was all that was needed. I suspect that the issue was related to file upload being involved.
I have around 10 plugins installed.
I use a wordpress theme by sitemile.
The site works fine sometimes, but very often, if I were to go to the login or sign up page, it displays the 500 internal error. I want to set up logs for see to see whats happening but I am not sure how I can do that.
Also, sometimes the homepage of the website loads very funky but if i go to other pages, it may or may not load that page properly.
Not an internet issue. Cleared cache as well. Interestingly, I see different results in normal and incognitu browsers.
Do these internal errors occur most of the time when you have a lot of plugins? Is there some other source of problem that I can check to make sure everything in my site is alright?
Error screenshot attached.
Homepage of website
Depending on your webhost, you might already have more details in a log file. Often times with apache it's called "error_log" and it can be found in your root directory or the directory of the file causing the error.
It could be anything from access permissions or a security block. So it's going to be hard for someone to give you an answer on this without more information.
Things to try:
You can try disabling one plugin at a time until it works.
Look for an error_log (or similar) file
Double check all of the file permissions, chmod and chown
Contact your web host and see if they can provide any insight
If you ARE the web host (or manage the server yourself) then you can check the apache error logs (if enabled) and/or security logs on the server
Double check any errors on the .htaccess file
A script could be consuming too much resources or taking too long to load, you can narrow it down by process of elimination and then disable/patch the culprit
Usually when I get the url of a request i use Request.RawUrl.
This gives /default.aspx for example.
However recently my host changed something and now the name of the application directory is displayed as well so i get /appdirname/default.aspx.
Now why does it give me the directory of the application? It looks as if my website is a subapplication of another website. So when you go to mydomain.com the rawurl will be:
/appdirname/default.aspx
I believe each domain has it's own website defined in iis or am i mistaken.
I am not asking for a workaround, which should be pretty straightforward, I am asking why this is happening and how, what kind of IIS setup causes this to happen?
PS.
And the worst part is i had this issue with godaddy and i was happy my host didnt have it but now both hosts have the same problem.
The Request.RawUrl method returns everything after the domain declaration, so if your full url is:
http://www.yourdomain.com:8080/directory/Page.aspx
then the method will return
/directory/Page.aspx
That's all it does. That's all it claims to do. As you say, your hosting provider must have changed something, which is very naughty, and the workaround should be easy. There is a good chance that they have introduced some kind of url redirection, but the best way to find out is to get in touch with their helpdesk and ask them what is happening. I find that most successful hosting companies tend to respond in good time to this kind of question. Otherwise they tend to become formerly-successful hosting companies.
Ric Strahl has this to say about it: http://www.west-wind.com/weblog/posts/132081.aspx
I'm creating an ASP.NET application which uses Facebook Connect and fbml tags. It also uses the LinkedIn widget. When I run this app in any browser, there are no warnings and everything works. However, in IE, a message like this comes up:
Security Warning:
The current webpage is trying to open a site in your Trusted sites list. Do you want to allow this?
Current site:http://www.facebook.com
Trusted site:http://localhost
(same for LinkedIn.com). I know how to fix this from a client perspective and to stop the security warning showing up. However, is it possible to ensure this message doesn't come up as it could be off putting for users who don't know how to suppress this warning? I haven't tried uploading it to my webhost, so not sure if this message will appear for everyone in production. However, I always get it on my local machine.
(None of my pages use SSL, so I don't think that's the issue. I tried using FB's HTTPS urls but that didn't make a difference).
Thanks
I have come across the IE message many times. Whilst this might not be the case here I always check in Firebug to see if any requests are going to Https (using Net tab). If may be the case that something you are referencing is itself making a call to something else.
Often you get that message if you are serving an https page and then going to fetch an image over http.
Might not help but is the first thing I do in this situation.