TIdHTTP generate many temporary internet files when upload? - http

I have an application that upload many files on the internet with TIdHTTP Indy component. Sometimes happen that many of those remain on the disk like "Temporary Internet Files" in the Internet Explorer directory and occupy several gigabyte.
Has anyone encountered this behavior?
Should I change the component?
[EDIT]: The cause of this problem wosn't TIdHTTP, but TDownloadURL! (that I use for download from another server)

Related

Edits in .asp net website wont show

i've got an issue with website made in asp.net. A site is published and online, i've made some modifications, republished site on my computer and just uploaded a .aspx file into the server via ftp.
First time it seems to have worked after a while. But i've made a small error and want to edit it again, i did the same, but it wont change. Could it be that i need to wait some time before changes are seen? Or could it be that there needs to be a server restart or something?
If you've edited something in the aspx.cs page you will need to upload the bin directory to the remote site, or better still republish the whole site.
If it is a change to the .aspx, css or javasctipt file, the original will most likely be cached in your browser. Try a differrent browser brand or refreshing the page, ctrl-f5 does a complete refresh.
If this error was by any chance a CSS mistake, that can be easily fixed by adding a "?" at the end of the address since CSS files are normally stored in the cache of the browser and the ? tells the browser to update them. Same thing is true about JavaScripts which are kept in individual files
I'd recommend you to use the Visual Studio Publish Website under the Build instead of manually uploading the site over FTP. That built in publisher provides you many advantages of which one of them is the same issue you have faced. When you make a small change, fixing the error in host would be very faster by republishing the site that way rather than manually upping it over FTP.

Iframes containing phps files are automatically downloaded

I really don't understand this, I created several tutorials that contain iframes with phps files(examples of code). On my localhost server they were displayed fine, now that I've uploaded it to my webhost it starts downloading the iframes sources when I open the tutorials.
What can be the cause of this??

Why does Web developer built-in web server serve zero byte files suddenly?

Visual Web Developer 2010 Express. C#, MVC3.
Clicking F5 to debug.
It starts up the built-in ASP web server on http://localhost:50188/
The Output window tells me WebDev.WebServer40.EXE is loading loads of DLLs.
Up until yesterday it worked. Today all URLs give me a blank page!
All controllers (all that changed yesterday was one controller, and some of its views; but it was working yesterday after all those changes). Same results in two different browsers. Use a different port gives an error (Telling me that there is something listening on port 50188!!) No errors anywhere. Just 0 byte files received.
My question is What happened and how do I fix it?
More Info:
Rebooting the machine made no difference.
I also found the obj/Debug directory and deleted it. It got recreated next time I hit F5 to debug. Still exactly the same problem!
And I went back 24hrs, in git, and still the same problem. So I'm sure the problem is not being caused by any of my source files. (The .csproj file is in git too.)
Look for a file called app_offline.htm (in your web root directory). It is a zero byte file. If it exists then this is served instead of any of your content! (It is a great feature if you wanted to take your site down for maintenance - put a custom message in that file.)
The Fix: Simply delete it and your website starts working again!
It appears (and I'm not sure about this) that the file is put there automatically when both you and your website want to access the DB at the same time. It should be deleted again automatically. But I guess a crash of something might leave it behind.
(To be honest, I think it would have been much wiser to put some content in app_offline.htm, explaining what it is and why it was automatically created. Quietly creating a zero-byte file is a tad sadistic...)
More information here: Why does app_offline.htm keep appearing in my web project?
And here: http://www.daniweb.com/web-development/aspnet/threads/215912/why-app_offline.htm-is-created-automatically-whats-the-mystery#

Upload more than 2GB file in Asp.Net?

I am trying to upload a 2+ GB video file using FTP . When I click “Upload” button in browser control, the page is not getting post but the status in browser is DONE. The same works fine with video files less than 2GB.
I just added one page, with browser control and one button for posting the page.
Just clicked the button, to post the page with 2gb file.
The page was not posting to server.
split the file up into smaller chunks, FTP those, and then re-combine.
MSDN
edit : You can do files up to 2GB + with ASP.NET using a third party solution that overrides the built-in request checking. This will work in all versions of IIS except IIS 7 integrated mode, which has a hard 2GB limit.
Read this
Read this : http://www.webdavsystem.com/server/documentation/upload
There are limits imposed by browsers on the amount of data you can upload in a single request.
In IE I think it is 2gb.
What are you using?

Does Flex save stuff to the local cache on client computer?

My Flex app downloads a bunch of images. When the user closes the browser window (or navigates away from the page), those images shouldn't be saved in the local cache. HTTP headers for the SWF file tells browser to fetch it from the website every time -- that is, it shouldn't save the SWF file locally. I just want to make sure that any of the images that the Flex app downloads isn't saved locally.
I checked Windows XP & Flash 9 for this. And nothing was stored locally. I assume this is true across all platforms and all versions of Flash greater than 9?
Thanks!
I think that flash download contents using the underling browser so the browser caches contents for him. In my experience sometimes browsers fails to read HTTP infos for caches and uses data from cache even if the file on the server is newer. The only way that works for sure to disable caching is changing the name of the file everytime you download it.
For example you can add a random string or a timestamp to your request. image1.gif is image1.gif?nocache=1231231

Resources